[ 465.730663] env[62199]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62199) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.731040] env[62199]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62199) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.731040] env[62199]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62199) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 465.731367] env[62199]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 465.819778] env[62199]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62199) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 465.828762] env[62199]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.009s {{(pid=62199) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 465.872371] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] Creating reply queue: reply_2b94d7d99e464fa289aad055ffbc8a8b [ 465.882085] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] Expecting reply to msg e05310c71486476d8e981f90d579cea2 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 465.894981] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e05310c71486476d8e981f90d579cea2 [ 466.430692] env[62199]: INFO nova.virt.driver [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 466.501006] env[62199]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 466.501249] env[62199]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.001s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 466.501345] env[62199]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62199) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 469.582068] env[62199]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-4310ace3-e020-4fd1-9d6c-fb520a8395f2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.598383] env[62199]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62199) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 469.598554] env[62199]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-219758a8-0f5e-4a90-9861-79a9f8357cd4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.629514] env[62199]: INFO oslo_vmware.api [-] Successfully established new session; session ID is d44e9. [ 469.629679] env[62199]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.128s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 469.630283] env[62199]: INFO nova.virt.vmwareapi.driver [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] VMware vCenter version: 7.0.3 [ 469.633704] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783fde08-fc06-478a-a024-5f868b1922c7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.654950] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40c3b123-e17d-47c2-b4e5-c607e7e5b7cd {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.661080] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-151d6c03-1c73-4444-b3cc-a00df05ff93c {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.667778] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe023c9f-5807-432e-af21-883251f8d602 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.680709] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b155d536-14c3-475d-9103-1c7d08422a1e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.686815] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4363086f-26ff-4b6e-b1c3-e209ffa6b9d1 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.716935] env[62199]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-651cab17-294d-42e8-914a-744af1d72689 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 469.722294] env[62199]: DEBUG nova.virt.vmwareapi.driver [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] Extension org.openstack.compute already exists. {{(pid=62199) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:224}} [ 469.725002] env[62199]: INFO nova.compute.provider_config [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 469.725677] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] Expecting reply to msg a6e9af61e77d4645a69691aaa4f5d469 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 469.742855] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a6e9af61e77d4645a69691aaa4f5d469 [ 470.228632] env[62199]: DEBUG nova.context [None req-55a517be-bcb4-403d-954b-c51722ffcd3b None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),6ee56b1c-b69d-42a2-9bc2-c3a38ae2088d(cell1) {{(pid=62199) load_cells /opt/stack/nova/nova/context.py:464}} [ 470.230895] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.231159] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.231831] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.232298] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Acquiring lock "6ee56b1c-b69d-42a2-9bc2-c3a38ae2088d" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.232491] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Lock "6ee56b1c-b69d-42a2-9bc2-c3a38ae2088d" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.233659] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Lock "6ee56b1c-b69d-42a2-9bc2-c3a38ae2088d" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.255004] env[62199]: INFO dbcounter [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Registered counter for database nova_cell0 [ 470.263620] env[62199]: INFO dbcounter [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Registered counter for database nova_cell1 [ 470.266792] env[62199]: DEBUG oslo_db.sqlalchemy.engines [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62199) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.267158] env[62199]: DEBUG oslo_db.sqlalchemy.engines [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62199) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.272174] env[62199]: ERROR nova.db.main.api [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.272174] env[62199]: result = function(*args, **kwargs) [ 470.272174] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 470.272174] env[62199]: return func(*args, **kwargs) [ 470.272174] env[62199]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.272174] env[62199]: result = fn(*args, **kwargs) [ 470.272174] env[62199]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.272174] env[62199]: return f(*args, **kwargs) [ 470.272174] env[62199]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 470.272174] env[62199]: return db.service_get_minimum_version(context, binaries) [ 470.272174] env[62199]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.272174] env[62199]: _check_db_access() [ 470.272174] env[62199]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.272174] env[62199]: stacktrace = ''.join(traceback.format_stack()) [ 470.272174] env[62199]: [ 470.273160] env[62199]: ERROR nova.db.main.api [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.273160] env[62199]: result = function(*args, **kwargs) [ 470.273160] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 470.273160] env[62199]: return func(*args, **kwargs) [ 470.273160] env[62199]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.273160] env[62199]: result = fn(*args, **kwargs) [ 470.273160] env[62199]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.273160] env[62199]: return f(*args, **kwargs) [ 470.273160] env[62199]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 470.273160] env[62199]: return db.service_get_minimum_version(context, binaries) [ 470.273160] env[62199]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.273160] env[62199]: _check_db_access() [ 470.273160] env[62199]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.273160] env[62199]: stacktrace = ''.join(traceback.format_stack()) [ 470.273160] env[62199]: [ 470.273570] env[62199]: WARNING nova.objects.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 470.273665] env[62199]: WARNING nova.objects.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Failed to get minimum service version for cell 6ee56b1c-b69d-42a2-9bc2-c3a38ae2088d [ 470.274090] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Acquiring lock "singleton_lock" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 470.274254] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Acquired lock "singleton_lock" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 470.274496] env[62199]: DEBUG oslo_concurrency.lockutils [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Releasing lock "singleton_lock" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 470.274814] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Full set of CONF: {{(pid=62199) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 470.274960] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ******************************************************************************** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2600}} [ 470.275088] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] Configuration options gathered from: {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2601}} [ 470.275225] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2602}} [ 470.275412] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2603}} [ 470.275540] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ================================================================================ {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2605}} [ 470.275747] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] allow_resize_to_same_host = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.275921] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] arq_binding_timeout = 300 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.276077] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] backdoor_port = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.276212] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] backdoor_socket = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.276380] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] block_device_allocate_retries = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.276544] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] block_device_allocate_retries_interval = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.276715] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cert = self.pem {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.276885] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.277053] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute_monitors = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.277216] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] config_dir = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.277385] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] config_drive_format = iso9660 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.277517] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.277754] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] config_source = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.277939] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] console_host = devstack {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.278107] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] control_exchange = nova {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.278276] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cpu_allocation_ratio = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.278430] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] daemon = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.278596] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] debug = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.278755] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] default_access_ip_network_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.278923] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] default_availability_zone = nova {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.279081] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] default_ephemeral_format = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.279241] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] default_green_pool_size = 1000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.279483] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.279649] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] default_schedule_zone = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.279894] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] disk_allocation_ratio = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.280094] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] enable_new_services = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.280281] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] enabled_apis = ['osapi_compute'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.280449] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] enabled_ssl_apis = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.280613] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] flat_injected = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.280803] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] force_config_drive = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.280981] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] force_raw_images = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.281154] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] graceful_shutdown_timeout = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.281402] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] heal_instance_info_cache_interval = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.281633] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] host = cpu-1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.281809] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.281978] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.282142] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.282354] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.282521] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instance_build_timeout = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.282685] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instance_delete_interval = 300 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.282854] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instance_format = [instance: %(uuid)s] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.283024] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instance_name_template = instance-%08x {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.283188] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instance_usage_audit = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.283360] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instance_usage_audit_period = month {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.283527] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.283695] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.283862] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] internal_service_availability_zone = internal {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.284031] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] key = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.284201] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] live_migration_retry_count = 30 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.284369] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_color = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.284536] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_config_append = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.284704] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.284865] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_dir = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.285026] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.285154] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_options = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.285317] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_rotate_interval = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.285485] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_rotate_interval_type = days {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.285658] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] log_rotation_type = none {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.285778] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.285904] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.286070] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.286234] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.286360] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.286523] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] long_rpc_timeout = 1800 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.286681] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] max_concurrent_builds = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.286841] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] max_concurrent_live_migrations = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.287001] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] max_concurrent_snapshots = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.287160] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] max_local_block_devices = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.287318] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] max_logfile_count = 30 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.287477] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] max_logfile_size_mb = 200 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.287702] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] maximum_instance_delete_attempts = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.287893] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] metadata_listen = 0.0.0.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.288096] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] metadata_listen_port = 8775 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.288281] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] metadata_workers = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.288452] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] migrate_max_retries = -1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.288617] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] mkisofs_cmd = genisoimage {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.288824] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.288960] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] my_ip = 10.180.1.21 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.289127] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] network_allocate_retries = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.289307] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.289476] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.289641] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] osapi_compute_listen_port = 8774 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.289885] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] osapi_compute_unique_server_name_scope = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.290085] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] osapi_compute_workers = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.290256] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] password_length = 12 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.290422] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] periodic_enable = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.290586] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] periodic_fuzzy_delay = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.290785] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] pointer_model = usbtablet {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.290975] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] preallocate_images = none {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.291142] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] publish_errors = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.291275] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] pybasedir = /opt/stack/nova {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.291524] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ram_allocation_ratio = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.291702] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] rate_limit_burst = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.291881] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] rate_limit_except_level = CRITICAL {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.292068] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] rate_limit_interval = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.292240] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] reboot_timeout = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.292402] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] reclaim_instance_interval = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.292561] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] record = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.292732] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] reimage_timeout_per_gb = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.292901] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] report_interval = 120 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.293064] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] rescue_timeout = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.293226] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] reserved_host_cpus = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.293387] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] reserved_host_disk_mb = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.293547] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] reserved_host_memory_mb = 512 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.293712] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] reserved_huge_pages = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.293880] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] resize_confirm_window = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.294043] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] resize_fs_using_block_device = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.294204] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] resume_guests_state_on_host_boot = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.294373] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.294536] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] rpc_response_timeout = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.294697] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] run_external_periodic_tasks = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.294865] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] running_deleted_instance_action = reap {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.295029] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.295189] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] running_deleted_instance_timeout = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.295349] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler_instance_sync_interval = 120 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.295519] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_down_time = 720 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.295687] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] servicegroup_driver = db {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.295848] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] shelved_offload_time = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.296014] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] shelved_poll_interval = 3600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.296220] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] shutdown_timeout = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.296385] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] source_is_ipv6 = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.296545] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ssl_only = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.296788] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.296959] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] sync_power_state_interval = 600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.297122] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] sync_power_state_pool_size = 1000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.297289] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] syslog_log_facility = LOG_USER {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.297448] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] tempdir = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.297664] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] timeout_nbd = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.297891] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] transport_url = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.298072] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] update_resources_interval = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.298237] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] use_cow_images = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.298396] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] use_eventlog = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.298557] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] use_journal = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.298718] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] use_json = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.298882] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] use_rootwrap_daemon = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.299043] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] use_stderr = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.299203] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] use_syslog = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.299362] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vcpu_pin_set = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.299530] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plugging_is_fatal = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.299700] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plugging_timeout = 300 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.299972] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] virt_mkfs = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.300179] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] volume_usage_poll_interval = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.300348] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] watch_log_file = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.300521] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] web = /usr/share/spice-html5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 470.300724] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_concurrency.disable_process_locking = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.301031] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.301216] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.301433] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.301653] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.301841] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.302016] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.302203] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.auth_strategy = keystone {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.302375] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.compute_link_prefix = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.302552] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.302730] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.dhcp_domain = novalocal {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.302904] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.enable_instance_password = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.303072] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.glance_link_prefix = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.303244] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.303421] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.303589] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.instance_list_per_project_cells = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.303753] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.list_records_by_skipping_down_cells = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.303963] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.local_metadata_per_cell = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.304174] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.max_limit = 1000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.304354] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.metadata_cache_expiration = 15 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.304534] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.neutron_default_tenant_id = default {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.304709] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.response_validation = warn {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.304884] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.use_neutron_default_nets = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.305070] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.305239] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.305411] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.305590] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.305766] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.vendordata_dynamic_targets = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.305935] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.vendordata_jsonfile_path = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.306118] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.306314] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.backend = dogpile.cache.memcached {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.306483] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.backend_argument = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.306656] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.config_prefix = cache.oslo {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.306829] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.dead_timeout = 60.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.306998] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.debug_cache_backend = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.307163] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.enable_retry_client = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.307957] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.enable_socket_keepalive = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.308198] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.enabled = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.308386] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.enforce_fips_mode = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.308564] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.expiration_time = 600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.308737] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.hashclient_retry_attempts = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.308912] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.309083] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_dead_retry = 300 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.309247] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_password = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.309416] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.309584] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.309801] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_pool_maxsize = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.309987] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.310161] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_sasl_enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.310346] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.310520] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.310686] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.memcache_username = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.310898] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.proxies = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.311079] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.redis_db = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.311244] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.redis_password = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.311421] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.311682] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.311875] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.redis_server = localhost:6379 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.312075] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.redis_socket_timeout = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.312262] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.redis_username = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.312436] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.retry_attempts = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.312606] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.retry_delay = 0.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.312776] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.socket_keepalive_count = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.312945] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.socket_keepalive_idle = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.313111] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.socket_keepalive_interval = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.313275] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.tls_allowed_ciphers = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.313435] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.tls_cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.313597] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.tls_certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.313763] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.tls_enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.313926] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cache.tls_keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.314099] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.314277] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.auth_type = password {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.314441] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.314620] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.314785] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.314952] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.315117] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.cross_az_attach = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.315281] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.debug = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.315442] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.endpoint_template = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.315609] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.http_retries = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.315774] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.315937] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.316135] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.os_region_name = RegionOne {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.316310] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.316474] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cinder.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.316651] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.316816] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.cpu_dedicated_set = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.316980] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.cpu_shared_set = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.317148] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.image_type_exclude_list = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.317317] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.317484] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.317653] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.317817] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.317991] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.318158] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.resource_provider_association_refresh = 300 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.318324] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.318489] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.shutdown_retry_interval = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.318670] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.318854] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] conductor.workers = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.319036] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] console.allowed_origins = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.319197] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] console.ssl_ciphers = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.319368] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] console.ssl_minimum_version = default {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.319539] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] consoleauth.enforce_session_timeout = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.319715] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] consoleauth.token_ttl = 600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.319909] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.320096] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.320271] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.320433] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.connect_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.320596] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.connect_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.320766] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.endpoint_override = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.320956] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.321120] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.321283] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.max_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.321441] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.min_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.321615] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.region_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.321872] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.retriable_status_codes = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.322052] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.service_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.322233] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.service_type = accelerator {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.322400] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.322565] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.status_code_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.322725] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.status_code_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.322888] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.323072] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.323242] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] cyborg.version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.323435] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.backend = sqlalchemy {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.323609] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.connection = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.323785] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.connection_debug = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.323981] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.connection_parameters = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.324170] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.connection_recycle_time = 3600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.324337] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.connection_trace = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.324502] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.db_inc_retry_interval = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.324669] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.db_max_retries = 20 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.324835] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.db_max_retry_interval = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.325002] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.db_retry_interval = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.325165] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.max_overflow = 50 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.325329] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.max_pool_size = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.325491] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.max_retries = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.325662] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.325825] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.mysql_wsrep_sync_wait = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.325987] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.pool_timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.326151] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.retry_interval = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.326310] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.slave_connection = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.326472] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.sqlite_synchronous = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.326633] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] database.use_db_reconnect = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.326822] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.backend = sqlalchemy {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.327014] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.connection = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.327188] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.connection_debug = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.327359] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.connection_parameters = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.327527] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.connection_recycle_time = 3600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.327692] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.connection_trace = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.327855] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.db_inc_retry_interval = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.328055] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.db_max_retries = 20 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.328235] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.db_max_retry_interval = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.328402] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.db_retry_interval = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.328566] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.max_overflow = 50 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.328746] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.max_pool_size = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.328884] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.max_retries = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.329058] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.329222] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.329384] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.pool_timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.329549] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.retry_interval = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.329719] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.slave_connection = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.329915] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] api_database.sqlite_synchronous = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.330103] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] devices.enabled_mdev_types = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.330283] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.330456] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.330623] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ephemeral_storage_encryption.enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.330789] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.330962] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.api_servers = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.331128] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.331290] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.331455] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.331615] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.connect_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.331848] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.connect_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.332041] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.debug = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.332218] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.default_trusted_certificate_ids = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.332381] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.enable_certificate_validation = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.332545] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.enable_rbd_download = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.332707] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.endpoint_override = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.332887] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.333072] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.333236] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.max_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.333397] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.min_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.333563] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.num_retries = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.333736] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.rbd_ceph_conf = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.333908] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.rbd_connect_timeout = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.334080] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.rbd_pool = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.334249] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.rbd_user = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.334414] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.region_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.334578] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.retriable_status_codes = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.334739] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.service_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.334913] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.service_type = image {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.335077] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.335237] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.status_code_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.335395] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.status_code_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.335554] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.335733] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.335929] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.verify_glance_signatures = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.336132] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] glance.version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.336312] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] guestfs.debug = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.336481] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] mks.enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.336841] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.337037] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] image_cache.manager_interval = 2400 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.337212] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] image_cache.precache_concurrency = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.337385] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] image_cache.remove_unused_base_images = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.337558] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.337729] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.337911] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] image_cache.subdirectory_name = _base {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.338090] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.api_max_retries = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.338261] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.api_retry_interval = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.338424] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.338589] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.auth_type = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.338752] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.338929] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.339116] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.339285] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.conductor_group = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.339449] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.connect_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.339611] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.connect_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.339800] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.endpoint_override = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.339980] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.340160] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.340324] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.max_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.340485] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.min_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.340653] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.peer_list = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.340826] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.region_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.341006] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.retriable_status_codes = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.341176] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.serial_console_state_timeout = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.341339] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.service_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.341512] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.service_type = baremetal {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.341676] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.shard = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.341927] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.342136] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.status_code_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.342307] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.status_code_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.342471] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.342657] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.342824] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ironic.version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.343009] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.343184] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] key_manager.fixed_key = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.343369] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.343532] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.barbican_api_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.343694] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.barbican_endpoint = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.343867] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.barbican_endpoint_type = public {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.344052] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.barbican_region_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.344225] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.344388] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.344553] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.344716] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.344878] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.345062] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.number_of_retries = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.345237] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.retry_delay = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.345403] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.send_service_user_token = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.345568] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.345731] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.345897] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.verify_ssl = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.346058] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican.verify_ssl_path = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.346227] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.346391] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.auth_type = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.346552] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.346712] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.346877] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.347040] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.347200] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.347363] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.347524] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] barbican_service_user.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.347694] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.approle_role_id = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.347881] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.approle_secret_id = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.348071] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.kv_mountpoint = secret {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.348240] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.kv_path = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.348409] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.kv_version = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.348571] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.namespace = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.348734] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.root_token_id = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.348899] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.ssl_ca_crt_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.349067] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.timeout = 60.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.349231] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.use_ssl = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.349399] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.349568] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.349748] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.auth_type = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.349920] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.350081] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.350246] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.350406] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.connect_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.350565] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.connect_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.350725] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.endpoint_override = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.350916] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.351079] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.351239] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.max_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.351397] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.min_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.351555] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.region_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.351715] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.retriable_status_codes = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.351896] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.service_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.352157] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.service_type = identity {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.352335] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.352500] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.status_code_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.352663] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.status_code_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.352825] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.353010] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.353173] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] keystone.version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.353376] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.connection_uri = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.353541] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.cpu_mode = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.353710] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.353909] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.cpu_models = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.354091] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.cpu_power_governor_high = performance {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.354263] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.354428] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.cpu_power_management = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.354600] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.354767] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.device_detach_attempts = 8 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.354939] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.device_detach_timeout = 20 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.355107] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.disk_cachemodes = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.355266] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.disk_prefix = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.355431] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.enabled_perf_events = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.355597] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.file_backed_memory = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.355766] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.gid_maps = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.355929] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.hw_disk_discard = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.356108] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.hw_machine_type = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.356287] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.images_rbd_ceph_conf = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.356457] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.356623] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.356804] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.images_rbd_glance_store_name = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.356997] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.images_rbd_pool = rbd {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.357174] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.images_type = default {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.357337] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.images_volume_group = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.357502] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.inject_key = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.357669] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.inject_partition = -2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.357834] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.inject_password = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.357999] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.iscsi_iface = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.358162] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.iser_use_multipath = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.358328] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.358492] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.358658] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_downtime = 500 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.358822] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.358987] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.359145] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_inbound_addr = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.359307] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.359472] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.359640] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_scheme = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.359856] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_timeout_action = abort {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.360067] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_tunnelled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.360243] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_uri = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.360410] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.live_migration_with_native_tls = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.360574] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.max_queues = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.360757] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.361013] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.361184] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.nfs_mount_options = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.361842] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.362111] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.362310] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.362485] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.362658] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.362832] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.num_pcie_ports = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.363016] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.363187] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.pmem_namespaces = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.363353] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.quobyte_client_cfg = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.363667] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.363850] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.364061] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.364211] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.364378] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rbd_secret_uuid = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.364542] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rbd_user = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.364712] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.364892] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.365132] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rescue_image_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.365321] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rescue_kernel_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.365490] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rescue_ramdisk_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.365665] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.365831] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.rx_queue_size = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.366008] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.smbfs_mount_options = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.366294] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.366471] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.snapshot_compression = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.366639] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.snapshot_image_format = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.366862] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.367035] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.sparse_logical_volumes = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.367203] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.swtpm_enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.367376] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.swtpm_group = tss {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.367546] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.swtpm_user = tss {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.367718] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.sysinfo_serial = unique {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.367884] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.tb_cache_size = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.368074] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.tx_queue_size = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.368305] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.uid_maps = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.368483] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.use_virtio_for_bridges = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.368659] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.virt_type = kvm {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.368835] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.volume_clear = zero {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.369006] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.volume_clear_size = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.369175] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.volume_use_multipath = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.369348] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.vzstorage_cache_path = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.369525] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.369698] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.369902] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.370082] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.370366] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.370546] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.vzstorage_mount_user = stack {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.370727] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.370912] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.371088] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.auth_type = password {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.371255] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.371419] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.371587] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.371752] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.connect_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.371918] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.connect_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.372181] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.default_floating_pool = public {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.372376] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.endpoint_override = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.372550] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.extension_sync_interval = 600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.372719] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.http_retries = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.372889] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.373056] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.373220] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.max_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.373394] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.373558] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.min_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.373729] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.ovs_bridge = br-int {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.373901] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.physnets = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.374074] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.region_name = RegionOne {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.374239] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.retriable_status_codes = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.374411] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.service_metadata_proxy = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.374575] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.service_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.374745] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.service_type = network {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.374913] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.375076] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.status_code_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.375239] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.status_code_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.375401] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.375582] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.375746] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] neutron.version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.375923] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] notifications.bdms_in_notifications = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.376132] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] notifications.default_level = INFO {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.376317] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] notifications.notification_format = unversioned {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.376487] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] notifications.notify_on_state_change = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.376666] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.376846] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] pci.alias = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.377020] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] pci.device_spec = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.377188] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] pci.report_in_placement = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.377363] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.377536] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.auth_type = password {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.377727] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.377898] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.378061] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.378225] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.378387] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.connect_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.378546] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.connect_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.378704] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.default_domain_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.378864] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.default_domain_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.379023] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.domain_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.379180] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.domain_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.379339] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.endpoint_override = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.379498] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.379653] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.379839] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.max_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.380094] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.min_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.380198] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.password = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.380362] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.project_domain_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.380531] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.project_domain_name = Default {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.380700] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.project_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.380900] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.project_name = service {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.381075] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.region_name = RegionOne {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.381241] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.retriable_status_codes = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.381403] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.service_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.381573] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.service_type = placement {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.381743] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.381907] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.status_code_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.382068] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.status_code_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.382308] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.system_scope = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.382496] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.382663] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.trust_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.382827] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.user_domain_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.383000] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.user_domain_name = Default {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.383161] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.user_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.383334] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.username = nova {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.383517] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.383682] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] placement.version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.383864] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.cores = 20 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.384044] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.count_usage_from_placement = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.384223] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.384397] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.injected_file_content_bytes = 10240 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.384568] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.injected_file_path_length = 255 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.384736] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.injected_files = 5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.384907] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.instances = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.385076] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.key_pairs = 100 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.385245] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.metadata_items = 128 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.385413] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.ram = 51200 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.385578] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.recheck_quota = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.385748] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.server_group_members = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.385918] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] quota.server_groups = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.386091] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.386260] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.386424] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.image_metadata_prefilter = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.386588] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.386754] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.max_attempts = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.386921] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.max_placement_results = 1000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.387087] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.387250] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.387695] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.387898] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] scheduler.workers = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.388110] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.388298] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.388483] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.388660] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.388831] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.389002] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.389168] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.389358] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.389531] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.host_subset_size = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.389700] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.389891] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.390061] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.390230] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.isolated_hosts = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.390394] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.isolated_images = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.390560] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.390736] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.390931] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.391095] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.pci_in_placement = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.391259] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.391421] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.391583] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.391748] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.391916] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.392102] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.392280] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.track_instance_changes = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.392551] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.392734] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] metrics.required = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.392907] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] metrics.weight_multiplier = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.393074] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.393239] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] metrics.weight_setting = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.393568] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.393747] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] serial_console.enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.393929] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] serial_console.port_range = 10000:20000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.394105] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.394275] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.394444] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] serial_console.serialproxy_port = 6083 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.394615] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.394792] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.auth_type = password {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.394958] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.395120] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.395285] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.395449] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.395612] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.395787] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.send_service_user_token = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.395956] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.396146] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] service_user.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.396325] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.agent_enabled = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.396493] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.396810] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.397022] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.397197] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.html5proxy_port = 6082 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.397361] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.image_compression = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.397524] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.jpeg_compression = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.397708] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.playback_compression = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.397894] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.server_listen = 127.0.0.1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.398068] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.398232] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.streaming_mode = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.398393] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] spice.zlib_compression = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.398561] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] upgrade_levels.baseapi = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.398736] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] upgrade_levels.compute = auto {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.398901] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] upgrade_levels.conductor = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.399062] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] upgrade_levels.scheduler = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.399232] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.399396] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.399557] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.399721] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.399909] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.400101] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.400272] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.400438] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.400600] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vendordata_dynamic_auth.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.400797] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.api_retry_count = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.400972] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.ca_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.401148] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.401320] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.cluster_name = testcl1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.401489] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.connection_pool_size = 10 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.401651] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.console_delay_seconds = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.401824] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.datastore_regex = ^datastore.* {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.402039] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.402211] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.host_password = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.402381] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.host_port = 443 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.402644] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.host_username = administrator@vsphere.local {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.402825] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.insecure = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.402994] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.integration_bridge = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.403163] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.maximum_objects = 100 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.403322] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.pbm_default_policy = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.403486] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.pbm_enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.403646] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.pbm_wsdl_location = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.403814] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.403977] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.serial_port_proxy_uri = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.404152] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.serial_port_service_uri = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.404321] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.task_poll_interval = 0.5 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.404493] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.use_linked_clone = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.404664] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.vnc_keymap = en-us {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.404832] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.vnc_port = 5900 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.405001] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vmware.vnc_port_total = 10000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.405189] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.auth_schemes = ['none'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.405364] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.405662] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.405849] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.406024] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.novncproxy_port = 6080 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.406205] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.server_listen = 127.0.0.1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.406379] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.406542] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.vencrypt_ca_certs = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.406703] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.vencrypt_client_cert = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.406864] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vnc.vencrypt_client_key = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.407042] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.407208] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.disable_deep_image_inspection = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.407371] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.407534] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.407721] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.407901] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.disable_rootwrap = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.408089] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.enable_numa_live_migration = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.408262] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.408426] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.408590] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.408755] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.libvirt_disable_apic = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.408921] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.409088] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.409253] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.409418] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.409591] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.409776] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.409954] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.410120] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.410286] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.410454] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.410643] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.410837] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.client_socket_timeout = 900 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.411014] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.default_pool_size = 1000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.411182] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.keep_alive = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.411351] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.max_header_line = 16384 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.411516] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.411679] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.ssl_ca_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.411842] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.ssl_cert_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.412010] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.ssl_key_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.412204] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.tcp_keepidle = 600 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.412429] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.412623] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] zvm.ca_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.412808] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] zvm.cloud_connector_url = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.413105] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.413283] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] zvm.reachable_timeout = 300 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.413466] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.enforce_new_defaults = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.413637] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.enforce_scope = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.413814] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.policy_default_rule = default {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.413997] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.414172] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.policy_file = policy.yaml {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.414344] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.414505] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.414664] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.414824] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.414989] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.415157] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.415332] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.415507] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.connection_string = messaging:// {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.415674] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.enabled = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.415844] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.es_doc_type = notification {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.416057] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.es_scroll_size = 10000 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.416195] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.es_scroll_time = 2m {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.416358] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.filter_error_trace = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.416527] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.hmac_keys = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.416696] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.sentinel_service_name = mymaster {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.416863] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.socket_timeout = 0.1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.417027] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.trace_requests = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.417190] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler.trace_sqlalchemy = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.417367] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler_jaeger.process_tags = {} {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.417531] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler_jaeger.service_name_prefix = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.417720] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] profiler_otlp.service_name_prefix = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.417900] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] remote_debug.host = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.418064] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] remote_debug.port = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.418247] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.418413] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.418578] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.418743] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.418906] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.419067] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.419226] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.419387] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.419548] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.419745] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.419910] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.420113] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.420291] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.420461] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.420632] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.420820] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.420994] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.421172] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.421339] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.421503] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.421699] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.421885] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.422056] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.422226] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.422391] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.422555] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.422809] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.422984] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.423158] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.423336] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.ssl = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.423519] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.423692] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.423858] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.424054] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.424214] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.424380] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.424568] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.424739] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_notifications.retry = -1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.424925] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.425100] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.425275] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.auth_section = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.425437] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.auth_type = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.425597] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.cafile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.425757] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.certfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.425922] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.collect_timing = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.426081] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.connect_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.426240] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.connect_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.426397] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.endpoint_id = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.426555] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.endpoint_override = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.426718] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.insecure = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.426899] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.keyfile = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.427114] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.max_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.427287] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.min_version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.427450] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.region_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.427615] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.retriable_status_codes = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.427777] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.service_name = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.427943] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.service_type = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.428135] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.split_loggers = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.428306] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.status_code_retries = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.428469] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.status_code_retry_delay = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.428631] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.timeout = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.428791] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.valid_interfaces = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.428953] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_limit.version = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.429120] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_reports.file_event_handler = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.429288] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.429449] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] oslo_reports.log_dir = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.429622] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.429810] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.429982] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.430152] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.430319] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.430479] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.430650] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.430831] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_ovs_privileged.group = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.430999] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.431165] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.431330] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.431490] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] vif_plug_ovs_privileged.user = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.431662] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.431841] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.432026] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.432207] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.432379] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.432548] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.432791] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.432979] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.433160] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.433333] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_ovs.isolate_vif = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.433511] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.433682] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.433853] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.434025] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.434189] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_vif_ovs.per_port_bridge = False {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.434358] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] os_brick.lock_path = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.434527] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] privsep_osbrick.capabilities = [21] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.434687] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] privsep_osbrick.group = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.434848] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] privsep_osbrick.helper_command = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.435012] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.435178] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.435337] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] privsep_osbrick.user = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.435508] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.435668] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] nova_sys_admin.group = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.435830] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] nova_sys_admin.helper_command = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.435998] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.436187] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.436350] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] nova_sys_admin.user = None {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 470.436483] env[62199]: DEBUG oslo_service.service [None req-332c669c-d1a4-43d4-9a89-93adeac8a195 None None] ******************************************************************************** {{(pid=62199) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2624}} [ 470.436899] env[62199]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 470.437757] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 7ea68720a5ad4c99a8b9274330ecccfc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 470.446402] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7ea68720a5ad4c99a8b9274330ecccfc [ 470.940148] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Getting list of instances from cluster (obj){ [ 470.940148] env[62199]: value = "domain-c8" [ 470.940148] env[62199]: _type = "ClusterComputeResource" [ 470.940148] env[62199]: } {{(pid=62199) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 470.941247] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c04c9ce6-058b-47e6-8fc6-084799ac561f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.950046] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Got total of 0 instances {{(pid=62199) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 470.950605] env[62199]: WARNING nova.virt.vmwareapi.driver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 470.951101] env[62199]: INFO nova.virt.node [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Generated node identity 715b0b11-2ed6-42ab-bb78-843b3217a820 [ 470.951350] env[62199]: INFO nova.virt.node [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Wrote node identity 715b0b11-2ed6-42ab-bb78-843b3217a820 to /opt/stack/data/n-cpu-1/compute_id [ 470.951754] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 4cd0c342e2864dfcb2e69220921c918b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 470.964787] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4cd0c342e2864dfcb2e69220921c918b [ 471.454098] env[62199]: WARNING nova.compute.manager [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Compute nodes ['715b0b11-2ed6-42ab-bb78-843b3217a820'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 471.454819] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 39a9c2754fa54f568e3bc518c925cda1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 471.479143] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 39a9c2754fa54f568e3bc518c925cda1 [ 471.957723] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 7a1d7ea37c354a579493f632bbcc111c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 471.969525] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7a1d7ea37c354a579493f632bbcc111c [ 472.460456] env[62199]: INFO nova.compute.manager [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 472.461054] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg b791c862de7c49c3ab12ebc10f56b9a3 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 472.485758] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b791c862de7c49c3ab12ebc10f56b9a3 [ 472.964224] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 6ab0935f024848e28ede7df6d3611055 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 472.976758] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6ab0935f024848e28ede7df6d3611055 [ 473.466904] env[62199]: WARNING nova.compute.manager [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 473.467179] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 473.467407] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 473.467559] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 473.467766] env[62199]: DEBUG nova.compute.resource_tracker [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62199) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 473.468669] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331d7084-7ebd-4a50-b983-bdb8d5c0aaf0 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.476769] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bbcd6e6-765c-4385-81b4-3be344a2940c {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.491439] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882a607b-c8aa-424d-aabd-726554d71246 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.497666] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d891828a-7ce4-45ab-8523-42e290311645 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 473.526324] env[62199]: DEBUG nova.compute.resource_tracker [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181752MB free_disk=126GB free_vcpus=48 pci_devices=None {{(pid=62199) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 473.526497] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 473.526677] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 473.527030] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 9793cc13c49241d68de16892cc08a935 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 473.538344] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9793cc13c49241d68de16892cc08a935 [ 474.029635] env[62199]: WARNING nova.compute.resource_tracker [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] No compute node record for cpu-1:715b0b11-2ed6-42ab-bb78-843b3217a820: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 715b0b11-2ed6-42ab-bb78-843b3217a820 could not be found. [ 474.030955] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg b088365df8a1449a83c4a4180068c763 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 474.042657] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b088365df8a1449a83c4a4180068c763 [ 474.544025] env[62199]: INFO nova.compute.resource_tracker [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 715b0b11-2ed6-42ab-bb78-843b3217a820 [ 474.544477] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 731616b7288949ab9824dbff944bd0c0 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 474.555223] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 731616b7288949ab9824dbff944bd0c0 [ 475.047103] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 5939c1ce0727403da83a684755fca13f in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 475.067241] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5939c1ce0727403da83a684755fca13f [ 475.563637] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 78282f5b5d3a4df7a213d4ed7ed066a4 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 475.584238] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 78282f5b5d3a4df7a213d4ed7ed066a4 [ 476.066256] env[62199]: DEBUG nova.compute.resource_tracker [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62199) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 476.066674] env[62199]: DEBUG nova.compute.resource_tracker [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62199) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 476.224762] env[62199]: INFO nova.scheduler.client.report [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] [req-018df4ad-4103-4e45-b2b9-bc3672f29575] Created resource provider record via placement API for resource provider with UUID 715b0b11-2ed6-42ab-bb78-843b3217a820 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 476.242603] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a14b793c-3bf4-40b6-8c15-50a13366a087 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.250131] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b728a0f1-7ab6-416c-9ddf-57a18c00df44 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.279284] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f191dd8f-41a7-415b-b4e4-866729d03d2b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.286524] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e48ea207-83ad-493d-83e8-f6b299642c56 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.299185] env[62199]: DEBUG nova.compute.provider_tree [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Updating inventory in ProviderTree for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.299770] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 500a7551db0c4897ae02f9b0e4cbed6e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 476.308055] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 500a7551db0c4897ae02f9b0e4cbed6e [ 476.840867] env[62199]: DEBUG nova.scheduler.client.report [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Updated inventory for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 476.841121] env[62199]: DEBUG nova.compute.provider_tree [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Updating resource provider 715b0b11-2ed6-42ab-bb78-843b3217a820 generation from 0 to 1 during operation: update_inventory {{(pid=62199) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.841261] env[62199]: DEBUG nova.compute.provider_tree [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Updating inventory in ProviderTree for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 476.892745] env[62199]: DEBUG nova.compute.provider_tree [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Updating resource provider 715b0b11-2ed6-42ab-bb78-843b3217a820 generation from 1 to 2 during operation: update_traits {{(pid=62199) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 476.895091] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Expecting reply to msg 0632f573ae444f50bf36e92f86a1e6a8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 476.909853] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0632f573ae444f50bf36e92f86a1e6a8 [ 477.398275] env[62199]: DEBUG nova.compute.resource_tracker [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62199) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 477.398626] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.872s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.398683] env[62199]: DEBUG nova.service [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Creating RPC server for service compute {{(pid=62199) start /opt/stack/nova/nova/service.py:186}} [ 477.410889] env[62199]: INFO oslo.messaging._drivers.impl_rabbit [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Creating fanout queue: compute_fanout_bb57d367976e40dbb8970948d0c5fde6 [ 477.414354] env[62199]: DEBUG nova.service [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] Join ServiceGroup membership for this service compute {{(pid=62199) start /opt/stack/nova/nova/service.py:203}} [ 477.414529] env[62199]: DEBUG nova.servicegroup.drivers.db [None req-6d22a73b-e168-4780-8905-4fdf6b36e331 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62199) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 482.416700] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 1614f8533856424e86ac282a6ee0b455 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 482.431096] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1614f8533856424e86ac282a6ee0b455 [ 497.420609] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._sync_power_states {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 497.420609] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 9abd81321c094151802550eb6608ffce in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 497.428634] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9abd81321c094151802550eb6608ffce [ 497.922711] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Getting list of instances from cluster (obj){ [ 497.922711] env[62199]: value = "domain-c8" [ 497.922711] env[62199]: _type = "ClusterComputeResource" [ 497.922711] env[62199]: } {{(pid=62199) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 497.924022] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7586643-a9f4-4b43-be98-933da59667a8 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.933008] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Got total of 0 instances {{(pid=62199) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 497.933237] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 497.933537] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Getting list of instances from cluster (obj){ [ 497.933537] env[62199]: value = "domain-c8" [ 497.933537] env[62199]: _type = "ClusterComputeResource" [ 497.933537] env[62199]: } {{(pid=62199) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 497.934368] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d52c1b49-c2a1-4d20-b91a-d407b606afa9 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 497.941733] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Got total of 0 instances {{(pid=62199) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 514.956349] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Acquiring lock "0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.956681] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Lock "0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.957038] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 1f93a781741e4689a255590695bdbc76 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 514.976035] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1f93a781741e4689a255590695bdbc76 [ 515.102425] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Acquiring lock "7384f309-7b32-40fc-881e-b967c7b8fec6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.102600] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Lock "7384f309-7b32-40fc-881e-b967c7b8fec6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.103112] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 5319dbdc2b1a4301ab058c7acabc5081 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 515.116271] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5319dbdc2b1a4301ab058c7acabc5081 [ 515.459237] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.461523] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 32e3983b13be416a87452ad24d1defcc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 515.530534] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquiring lock "47b420a2-5eeb-4f14-9137-fb591a9dcbe6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.530923] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Lock "47b420a2-5eeb-4f14-9137-fb591a9dcbe6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.531306] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg 26c319bc21d444a48516088a4ca23b31 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 515.541247] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 32e3983b13be416a87452ad24d1defcc [ 515.544542] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 26c319bc21d444a48516088a4ca23b31 [ 515.605021] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 515.607143] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg f4c1c34d3cea4706aed2facaaa3a8c05 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 515.662345] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f4c1c34d3cea4706aed2facaaa3a8c05 [ 515.991840] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 515.992257] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 515.993802] env[62199]: INFO nova.compute.claims [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 515.995655] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 4a5e9f84d2f44d8e8931f612389821f0 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 516.036192] env[62199]: DEBUG nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.037861] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg 7f8aa31ce0a44946bd7dbc84fa2e0fb7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 516.050191] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4a5e9f84d2f44d8e8931f612389821f0 [ 516.120978] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7f8aa31ce0a44946bd7dbc84fa2e0fb7 [ 516.130595] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.409083] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Acquiring lock "e34b059b-8edf-4a81-9485-358cefd2bde6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.409356] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Lock "e34b059b-8edf-4a81-9485-358cefd2bde6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.409874] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 0cedac7da7f14ea9a112eac7c00740eb in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 516.425998] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0cedac7da7f14ea9a112eac7c00740eb [ 516.499334] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 819fd924112240038947cd45fb526a22 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 516.512116] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 819fd924112240038947cd45fb526a22 [ 516.556483] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.884151] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 516.884151] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.004s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 516.884151] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 8f212e6501894db39f2afb82668d6277 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 516.904761] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8f212e6501894db39f2afb82668d6277 [ 516.914172] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 516.915898] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg b9ff57be67884b0f84606dcd22d7240a in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 516.966922] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b9ff57be67884b0f84606dcd22d7240a [ 517.157484] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-970784cb-4ebe-4f0a-8b3a-f2332d298feb {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.166027] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9558acb-7496-4247-bd22-d4ae43b06402 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.204142] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65aeefc1-e2c7-4364-9e5c-4d3ab1bb75df {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.218419] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e8fbe8-4242-4dd7-bd63-2455afb7860e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 517.240699] env[62199]: DEBUG nova.compute.provider_tree [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 517.241385] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 5da3e2084a3c4b5c861a0161118a2e29 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 517.249762] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5da3e2084a3c4b5c861a0161118a2e29 [ 517.274008] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Acquiring lock "24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.274238] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Lock "24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.274692] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 3fd9fa71c3124b56bf7f02864c716d8e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 517.285175] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3fd9fa71c3124b56bf7f02864c716d8e [ 517.385866] env[62199]: DEBUG nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.388386] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg dbd7bb158ab54cf6bad44fab81749b14 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 517.438444] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg dbd7bb158ab54cf6bad44fab81749b14 [ 517.446461] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.746821] env[62199]: DEBUG nova.scheduler.client.report [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 517.749299] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg ae5fe274c1e445eb93445f5dc0fbe08f in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 517.771432] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ae5fe274c1e445eb93445f5dc0fbe08f [ 517.780035] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.781804] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 61d8ac046f5d4ab3bb843310492993d9 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 517.834158] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 61d8ac046f5d4ab3bb843310492993d9 [ 517.910018] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.255732] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.262s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 518.255732] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 518.256809] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg f67f1191e14e434f9f02348a712cd9c7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 518.259442] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.128s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.260099] env[62199]: INFO nova.compute.claims [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.271898] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 9cbe974596fc41be922625bd69d4fed9 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 518.302857] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f67f1191e14e434f9f02348a712cd9c7 [ 518.304168] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.338756] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9cbe974596fc41be922625bd69d4fed9 [ 518.377681] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Acquiring lock "52bd7ff6-1149-4a56-95c2-806572b526f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.378015] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Lock "52bd7ff6-1149-4a56-95c2-806572b526f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.378631] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 33aff6d4b71942df9dc545d49f90abca in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 518.391698] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 33aff6d4b71942df9dc545d49f90abca [ 518.782766] env[62199]: DEBUG nova.compute.utils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 518.782766] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg d837ee57121542ddb9765c6a2c2a7788 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 518.782766] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 518.782766] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 518.785676] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 567e40346d0743be8dd0a6e894ccb7b0 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 518.816227] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 567e40346d0743be8dd0a6e894ccb7b0 [ 518.817410] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d837ee57121542ddb9765c6a2c2a7788 [ 518.881736] env[62199]: DEBUG nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.884925] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 11ec2304bbc446deb3521b9974859804 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 518.962982] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 11ec2304bbc446deb3521b9974859804 [ 519.257327] env[62199]: DEBUG nova.policy [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8ea3ad252b51443fb05df4daaacd53d8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd6e245649d10465e9855306f241cbfa5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 519.284660] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 519.286414] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 6ed7616a8cfe461bbaf11a34badd6e93 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 519.346842] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6ed7616a8cfe461bbaf11a34badd6e93 [ 519.410315] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.434291] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d6d6a2-5799-4de8-ba09-0c4dd44e468d {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.442184] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31e29d35-54be-4d56-a82f-f8fa0cb73fa6 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.478866] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a78bdc-ab3f-453a-8d4d-5537279e0347 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.486962] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-217a99b8-866e-4558-906f-06c498de141c {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.504861] env[62199]: DEBUG nova.compute.provider_tree [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.505429] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 53a4f6f6be984f4d8d6c3020c6f973dd in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 519.512539] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 53a4f6f6be984f4d8d6c3020c6f973dd [ 519.795667] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 08013fdfe953477dbb0389dcee4830f5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 519.841680] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 08013fdfe953477dbb0389dcee4830f5 [ 519.868633] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Successfully created port: 9a060963-bba6-4aa8-968a-e6bef71637e0 {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 519.966681] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.966906] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Lock "a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 519.967343] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 0ad1dee6d2bb4a36b234ed4771030981 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 519.977041] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0ad1dee6d2bb4a36b234ed4771030981 [ 520.008554] env[62199]: DEBUG nova.scheduler.client.report [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.010944] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 714f30b6979341fcb78ccef223c4b6d2 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 520.031494] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 714f30b6979341fcb78ccef223c4b6d2 [ 520.298841] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 520.323296] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 520.323538] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 520.323691] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 520.323867] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 520.324027] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 520.324193] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 520.324401] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 520.324556] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 520.324930] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 520.325101] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 520.325269] env[62199]: DEBUG nova.virt.hardware [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 520.326118] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db6d2be8-5c7f-4f6c-8885-36571efc70ae {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.333797] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-730b5a03-8508-48fc-871e-28688224d3ee {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.349775] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c88cf9e2-d253-4fa3-bde7-d57b35a31d38 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.468989] env[62199]: DEBUG nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.470784] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 6451ec8d68304252993963d6d023a12a in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 520.518474] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.260s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.519039] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.520793] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 4e5e4e623eef4c5083306fc8d29f3c02 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 520.521782] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.965s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.523145] env[62199]: INFO nova.compute.claims [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.524643] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg ecb9b62b6b004746bcfa85f03aead0eb in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 520.536896] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6451ec8d68304252993963d6d023a12a [ 520.583131] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4e5e4e623eef4c5083306fc8d29f3c02 [ 520.588728] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ecb9b62b6b004746bcfa85f03aead0eb [ 520.991945] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.027947] env[62199]: DEBUG nova.compute.utils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.028666] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg a3b66d7d1eed4f448a65086fb4051e6f in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 521.033933] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg f8da13824b9d457f9c140feebcf3626b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 521.033933] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 521.033933] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 521.039084] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a3b66d7d1eed4f448a65086fb4051e6f [ 521.039880] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f8da13824b9d457f9c140feebcf3626b [ 521.090482] env[62199]: DEBUG nova.policy [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf4be925fb5e4f50a1e1914bbb78b81d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '536159c70ea64ac1851cc2990162c4a1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 521.535466] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.535466] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg c46222a81da3471991ad85e0ddeee518 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 521.572204] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c46222a81da3471991ad85e0ddeee518 [ 521.667700] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad96cbe1-a353-47b6-8b89-553003295741 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.676114] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66b02057-3b57-4769-b24c-27f54960dbad {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.709627] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ab3886-cb9a-4cbe-983d-9660b2309d30 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.723300] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c67d791-98ed-43bc-ba8e-c7b1bea292e0 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.735153] env[62199]: DEBUG nova.compute.provider_tree [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.735701] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg 6ac5bdcf8982454faccbd2c640ed53be in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 521.748998] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6ac5bdcf8982454faccbd2c640ed53be [ 521.800497] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Successfully created port: 4715eda5-a164-4e5c-9a61-c2317e4046b6 {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 522.039799] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 386965ea4d9242aab666dfcf9e7fac0a in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 522.073301] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 386965ea4d9242aab666dfcf9e7fac0a [ 522.238505] env[62199]: DEBUG nova.scheduler.client.report [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.240895] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg ee1705017f164d19b44e4302c6831d83 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 522.256259] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ee1705017f164d19b44e4302c6831d83 [ 522.535421] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "2681b401-47b3-4b05-b923-65e3078074d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.535648] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Lock "2681b401-47b3-4b05-b923-65e3078074d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.536122] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 1ae2d3c01a6646bc8a09139a8dc1cd69 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 522.543450] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.546486] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1ae2d3c01a6646bc8a09139a8dc1cd69 [ 522.565109] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.565555] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.565935] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.566356] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.566670] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.566921] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.567322] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.567575] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.567891] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.568201] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.568484] env[62199]: DEBUG nova.virt.hardware [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.569724] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8439d1-ceb5-4242-a2e0-29c46160871c {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.580231] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bda25464-0b19-4fc9-85f9-4aa52bc5ad4b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.752732] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.231s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.753321] env[62199]: DEBUG nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.758201] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg a72966e6a2014557a63b9832f05199f5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 522.758201] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.310s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.758201] env[62199]: INFO nova.compute.claims [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 522.759048] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 757096f68b794f4484fe8b0de374134b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 522.805282] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a72966e6a2014557a63b9832f05199f5 [ 522.813001] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 757096f68b794f4484fe8b0de374134b [ 523.037833] env[62199]: DEBUG nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 523.039582] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 22722e39c97846769bc74246078f68c4 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 523.076719] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 22722e39c97846769bc74246078f68c4 [ 523.263017] env[62199]: DEBUG nova.compute.utils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.264118] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg d61469fe044949eda411366a3432b911 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 523.265928] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 330c3436537b4b7daccf633e78956c75 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 523.266876] env[62199]: DEBUG nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Not allocating networking since 'none' was specified. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 523.276724] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 330c3436537b4b7daccf633e78956c75 [ 523.280920] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d61469fe044949eda411366a3432b911 [ 523.447816] env[62199]: ERROR nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. [ 523.447816] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.447816] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.447816] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.447816] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.447816] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.447816] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.447816] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.447816] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.447816] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 523.447816] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.447816] env[62199]: ERROR nova.compute.manager raise self.value [ 523.447816] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.447816] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.447816] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.447816] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.448421] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.448421] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.448421] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. [ 523.448421] env[62199]: ERROR nova.compute.manager [ 523.448653] env[62199]: Traceback (most recent call last): [ 523.448653] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.448653] env[62199]: listener.cb(fileno) [ 523.448653] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.448653] env[62199]: result = function(*args, **kwargs) [ 523.448653] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.448653] env[62199]: return func(*args, **kwargs) [ 523.448653] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.448653] env[62199]: raise e [ 523.448653] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.448653] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 523.448653] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.448653] env[62199]: created_port_ids = self._update_ports_for_instance( [ 523.448653] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.448653] env[62199]: with excutils.save_and_reraise_exception(): [ 523.448653] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.448653] env[62199]: self.force_reraise() [ 523.448653] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.448653] env[62199]: raise self.value [ 523.448653] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.448653] env[62199]: updated_port = self._update_port( [ 523.448653] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.448653] env[62199]: _ensure_no_port_binding_failure(port) [ 523.448653] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.448653] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.448653] env[62199]: nova.exception.PortBindingFailed: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. [ 523.448653] env[62199]: Removing descriptor: 15 [ 523.450188] env[62199]: ERROR nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Traceback (most recent call last): [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] yield resources [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self.driver.spawn(context, instance, image_meta, [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] vm_ref = self.build_virtual_machine(instance, [ 523.450188] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] for vif in network_info: [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return self._sync_wrapper(fn, *args, **kwargs) [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self.wait() [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self[:] = self._gt.wait() [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return self._exit_event.wait() [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.450578] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] result = hub.switch() [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return self.greenlet.switch() [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] result = function(*args, **kwargs) [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return func(*args, **kwargs) [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] raise e [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] nwinfo = self.network_api.allocate_for_instance( [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] created_port_ids = self._update_ports_for_instance( [ 523.450942] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] with excutils.save_and_reraise_exception(): [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self.force_reraise() [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] raise self.value [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] updated_port = self._update_port( [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] _ensure_no_port_binding_failure(port) [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] raise exception.PortBindingFailed(port_id=port['id']) [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] nova.exception.PortBindingFailed: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. [ 523.451267] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] [ 523.451591] env[62199]: INFO nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Terminating instance [ 523.453324] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Acquiring lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.453492] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Acquired lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.453660] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 523.454077] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 878ff33569be4fd4b7f24bd312f69e3c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 523.468783] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 878ff33569be4fd4b7f24bd312f69e3c [ 523.564423] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.767522] env[62199]: DEBUG nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.769614] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg 28dfa0b75a994dbaa4ba83d6f9afabcd in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 523.817696] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 28dfa0b75a994dbaa4ba83d6f9afabcd [ 523.928607] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fd0ab5-4047-4a90-a665-94a74377d10b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.941524] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cfe372e-eec5-4732-a2ff-fe985a5f3650 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.984850] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ab62d8-d9ad-4c2a-9722-17da6c7b5f85 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.992833] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8a3c6f-8a57-46b8-b7b5-b7f888bf1d6d {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.008103] env[62199]: DEBUG nova.compute.provider_tree [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.008103] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg f82cb25fb91c445089808bdbb3a99d69 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 524.014810] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f82cb25fb91c445089808bdbb3a99d69 [ 524.129764] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.267360] env[62199]: DEBUG nova.compute.manager [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Received event network-changed-9a060963-bba6-4aa8-968a-e6bef71637e0 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 524.267564] env[62199]: DEBUG nova.compute.manager [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Refreshing instance network info cache due to event network-changed-9a060963-bba6-4aa8-968a-e6bef71637e0. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 524.268225] env[62199]: DEBUG oslo_concurrency.lockutils [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] Acquiring lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.270097] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.270547] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg f26f4acec3f04ca48cd18f88568501fe in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 524.277184] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg 5f97a138361544589439ab737f2524c8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 524.287584] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f26f4acec3f04ca48cd18f88568501fe [ 524.323916] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5f97a138361544589439ab737f2524c8 [ 524.516450] env[62199]: DEBUG nova.scheduler.client.report [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.516450] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg f390a52b3a6f4a47a4f63d1b6cb0d5a6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 524.528602] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f390a52b3a6f4a47a4f63d1b6cb0d5a6 [ 524.667210] env[62199]: ERROR nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. [ 524.667210] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 524.667210] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.667210] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 524.667210] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.667210] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 524.667210] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.667210] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 524.667210] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.667210] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 524.667210] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.667210] env[62199]: ERROR nova.compute.manager raise self.value [ 524.667210] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.667210] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 524.667210] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.667210] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 524.667868] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.667868] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 524.667868] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. [ 524.667868] env[62199]: ERROR nova.compute.manager [ 524.667868] env[62199]: Traceback (most recent call last): [ 524.667868] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 524.667868] env[62199]: listener.cb(fileno) [ 524.667868] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.667868] env[62199]: result = function(*args, **kwargs) [ 524.667868] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.667868] env[62199]: return func(*args, **kwargs) [ 524.667868] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.667868] env[62199]: raise e [ 524.667868] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.667868] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 524.667868] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.667868] env[62199]: created_port_ids = self._update_ports_for_instance( [ 524.667868] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.667868] env[62199]: with excutils.save_and_reraise_exception(): [ 524.667868] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.667868] env[62199]: self.force_reraise() [ 524.667868] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.667868] env[62199]: raise self.value [ 524.667868] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.667868] env[62199]: updated_port = self._update_port( [ 524.667868] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.667868] env[62199]: _ensure_no_port_binding_failure(port) [ 524.667868] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.667868] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 524.668711] env[62199]: nova.exception.PortBindingFailed: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. [ 524.668711] env[62199]: Removing descriptor: 16 [ 524.668711] env[62199]: ERROR nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Traceback (most recent call last): [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] yield resources [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self.driver.spawn(context, instance, image_meta, [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 524.668711] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] vm_ref = self.build_virtual_machine(instance, [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] vif_infos = vmwarevif.get_vif_info(self._session, [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] for vif in network_info: [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return self._sync_wrapper(fn, *args, **kwargs) [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self.wait() [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self[:] = self._gt.wait() [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return self._exit_event.wait() [ 524.669020] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] result = hub.switch() [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return self.greenlet.switch() [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] result = function(*args, **kwargs) [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return func(*args, **kwargs) [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] raise e [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] nwinfo = self.network_api.allocate_for_instance( [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 524.669354] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] created_port_ids = self._update_ports_for_instance( [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] with excutils.save_and_reraise_exception(): [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self.force_reraise() [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] raise self.value [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] updated_port = self._update_port( [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] _ensure_no_port_binding_failure(port) [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 524.669696] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] raise exception.PortBindingFailed(port_id=port['id']) [ 524.670063] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] nova.exception.PortBindingFailed: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. [ 524.670063] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] [ 524.670063] env[62199]: INFO nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Terminating instance [ 524.671048] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Acquiring lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 524.671212] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Acquired lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.671372] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 524.671783] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg a8ff1f8ab72c469a81ab5e2ea692381a in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 524.680806] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a8ff1f8ab72c469a81ab5e2ea692381a [ 524.773010] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Releasing lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.773623] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.773623] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 524.774561] env[62199]: DEBUG oslo_concurrency.lockutils [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] Acquired lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 524.774561] env[62199]: DEBUG nova.network.neutron [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Refreshing network info cache for port 9a060963-bba6-4aa8-968a-e6bef71637e0 {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 524.774561] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] Expecting reply to msg e53b32d9f8c242178f9f883d974279ea in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 524.775924] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-16900191-6c48-46cd-b9aa-7fc232abc6e1 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.781196] env[62199]: DEBUG nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.783320] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e53b32d9f8c242178f9f883d974279ea [ 524.787893] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8f9b06b-5521-4c3b-a30f-1b9ec5be8c79 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.811081] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3 could not be found. [ 524.811303] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 524.812039] env[62199]: INFO nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 524.812039] env[62199]: DEBUG oslo.service.loopingcall [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.812247] env[62199]: DEBUG nova.compute.manager [-] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.812338] env[62199]: DEBUG nova.network.neutron [-] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 524.836986] env[62199]: DEBUG nova.network.neutron [-] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 524.837537] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg ea8322335cb942d69ddb6a2eb5bffd3d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 524.846828] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ea8322335cb942d69ddb6a2eb5bffd3d [ 525.019029] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.263s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.019230] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.021040] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 4b2bc4c51d6c427da84142e0c26041ff in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.022157] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.112s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.023630] env[62199]: INFO nova.compute.claims [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.025125] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 3da2a09d87594c45bcbc10ebb2574ee2 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.065894] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4b2bc4c51d6c427da84142e0c26041ff [ 525.067253] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3da2a09d87594c45bcbc10ebb2574ee2 [ 525.237287] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.345002] env[62199]: DEBUG nova.network.neutron [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 525.347098] env[62199]: DEBUG nova.network.neutron [-] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.347562] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg ddc61a3f9c624567926061852e8368c6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.357693] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ddc61a3f9c624567926061852e8368c6 [ 525.425575] env[62199]: DEBUG nova.network.neutron [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.426215] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] Expecting reply to msg a983b56ae95d4bc598e341c85252d202 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.437252] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a983b56ae95d4bc598e341c85252d202 [ 525.466609] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 525.466609] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 525.466609] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 525.466609] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 525.466841] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 525.466879] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 525.467336] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 525.467336] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 525.467452] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 525.467550] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 525.467697] env[62199]: DEBUG nova.virt.hardware [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 525.468639] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee3acff1-3a54-4005-9ed6-0a89107c2549 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.479589] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f90d6a39-cb43-40b0-b862-3a486cdf8ea3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.499036] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Instance VIF info [] {{(pid=62199) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 525.515827] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62199) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.515827] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-21e5e7b6-3610-4a4c-b9b7-21d389665bf4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.526221] env[62199]: INFO nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Created folder: OpenStack in parent group-v4. [ 525.526393] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating folder: Project (9f6e3bfb3216497989cba9b77321059b). Parent ref: group-v108209. {{(pid=62199) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.528024] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg e066e4dc327a4417aef265c03c43dad6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.528848] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d0785543-65e5-43d5-95d7-9735f5828c90 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.531551] env[62199]: DEBUG nova.compute.utils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.532137] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg f3ead55b671c4a119adab46308eb57f6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.537022] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.537022] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 525.542737] env[62199]: INFO nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Created folder: Project (9f6e3bfb3216497989cba9b77321059b) in parent group-v108209. [ 525.542914] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating folder: Instances. Parent ref: group-v108210. {{(pid=62199) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 525.543132] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-039844db-bc45-4f6a-a7a1-c789c0a5fd4f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.545572] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e066e4dc327a4417aef265c03c43dad6 [ 525.546127] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f3ead55b671c4a119adab46308eb57f6 [ 525.555109] env[62199]: INFO nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Created folder: Instances in parent group-v108210. [ 525.555180] env[62199]: DEBUG oslo.service.loopingcall [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 525.555631] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Creating VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 525.555631] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f08cf1a3-ff34-48a0-981f-75cfbd0470df {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.574198] env[62199]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 525.574198] env[62199]: value = "task-398546" [ 525.574198] env[62199]: _type = "Task" [ 525.574198] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 525.583722] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398546, 'name': CreateVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 525.802263] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.802263] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 1daa6f4ca0cc4b5b86cd54c2dc23325d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.812111] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1daa6f4ca0cc4b5b86cd54c2dc23325d [ 525.850241] env[62199]: INFO nova.compute.manager [-] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Took 1.04 seconds to deallocate network for instance. [ 525.853031] env[62199]: DEBUG nova.compute.claims [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 525.853215] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.880283] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.880597] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 525.880792] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Starting heal instance info cache {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9926}} [ 525.880888] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Rebuilding the list of instances to heal {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9930}} [ 525.881489] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg fd5ab61082034ade84a046cf500d9d00 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 525.896384] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg fd5ab61082034ade84a046cf500d9d00 [ 525.928647] env[62199]: DEBUG oslo_concurrency.lockutils [req-dc4e578b-05d0-4fa6-967b-100bf915f655 req-ef393e1a-39f4-4da1-8203-9040f52d9086 service nova] Releasing lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.950392] env[62199]: DEBUG nova.policy [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '369fa05c3e134da9bb74cd29240936e6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'de94d9183d824324a17d9f66d275d1a9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 526.043406] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.043406] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 441e32a9bab2463ab806826c94525ea8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 526.084190] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398546, 'name': CreateVM_Task} progress is 25%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.084919] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 441e32a9bab2463ab806826c94525ea8 [ 526.216266] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3972d5c-19f0-4e06-8a5e-1273b2683744 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.223940] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33301530-690a-44ff-b9a0-363682d2677e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.254718] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cead47b-58b7-4fc5-bb3e-e8d73a81f7b1 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.275361] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a97cfa-682e-49c3-a413-2e98edcfc9ad {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.290786] env[62199]: DEBUG nova.compute.provider_tree [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.291380] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 0f4b9bc249734129b422ee9de7424121 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 526.298288] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0f4b9bc249734129b422ee9de7424121 [ 526.303964] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Releasing lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.304651] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 526.304850] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 526.305143] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5bc09c28-7410-424d-9bf4-420f8fc0ac34 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.315469] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06295eca-e55f-4a2d-989f-d22586699002 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.342559] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7384f309-7b32-40fc-881e-b967c7b8fec6 could not be found. [ 526.342802] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 526.343009] env[62199]: INFO nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 526.343212] env[62199]: DEBUG oslo.service.loopingcall [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 526.343484] env[62199]: DEBUG nova.compute.manager [-] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 526.343613] env[62199]: DEBUG nova.network.neutron [-] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 526.385701] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Skipping network cache update for instance because it is Building. {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 526.385950] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Skipping network cache update for instance because it is Building. {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 526.386098] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Skipping network cache update for instance because it is Building. {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 526.386225] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Skipping network cache update for instance because it is Building. {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 526.386414] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Skipping network cache update for instance because it is Building. {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9939}} [ 526.386550] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Didn't find any instances for network info cache update. {{(pid=62199) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10012}} [ 526.386739] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.387236] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.387446] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.387614] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.387787] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.388057] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.388776] env[62199]: DEBUG nova.compute.manager [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62199) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10545}} [ 526.388776] env[62199]: DEBUG oslo_service.periodic_task [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Running periodic task ComputeManager.update_available_resource {{(pid=62199) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.389241] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 7a04e3c33f38487b935c6b8c66b64c86 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 526.399919] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7a04e3c33f38487b935c6b8c66b64c86 [ 526.430320] env[62199]: DEBUG nova.network.neutron [-] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 526.430883] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 0a7728ac1cfb46e9be70a3b6c0202bef in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 526.439682] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0a7728ac1cfb46e9be70a3b6c0202bef [ 526.440289] env[62199]: DEBUG nova.network.neutron [-] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 526.441003] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 5f72323757d34281add76b1a4f67b4fd in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 526.450931] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5f72323757d34281add76b1a4f67b4fd [ 526.548710] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 5ec4fbc505744aa5b46fd2f8dd1e6aff in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 526.589708] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398546, 'name': CreateVM_Task} progress is 25%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 526.620173] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5ec4fbc505744aa5b46fd2f8dd1e6aff [ 526.797742] env[62199]: DEBUG nova.scheduler.client.report [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 526.797742] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 94b7b74e31d64903a6b7d42d21e0888c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 526.828457] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 94b7b74e31d64903a6b7d42d21e0888c [ 526.843216] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Successfully created port: da4c2f45-8714-4d2c-8997-38c3c011820b {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 526.896657] env[62199]: DEBUG oslo_concurrency.lockutils [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.948371] env[62199]: INFO nova.compute.manager [-] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Took 0.60 seconds to deallocate network for instance. [ 526.950733] env[62199]: DEBUG nova.compute.claims [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 526.951016] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.051797] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.076212] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.076520] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.076671] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.076864] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.077002] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.077139] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.077336] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.077485] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.077642] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.077793] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.077953] env[62199]: DEBUG nova.virt.hardware [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.078849] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917ecc23-13e0-4a08-8b46-8c6bf1a4f88e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.090565] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b678a0a1-9265-4c49-a5af-148835041908 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.110953] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398546, 'name': CreateVM_Task} progress is 99%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.303125] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.281s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.303633] env[62199]: DEBUG nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.305489] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 8c4a41c85f1b4d38add04ace7de9fb6e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 527.306329] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.002s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.307662] env[62199]: INFO nova.compute.claims [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.310282] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg c527ffb8f097455ab4444c6259e3ec20 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 527.329485] env[62199]: DEBUG nova.compute.manager [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Received event network-vif-deleted-9a060963-bba6-4aa8-968a-e6bef71637e0 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 527.329485] env[62199]: DEBUG nova.compute.manager [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Received event network-changed-4715eda5-a164-4e5c-9a61-c2317e4046b6 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 527.329485] env[62199]: DEBUG nova.compute.manager [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Refreshing instance network info cache due to event network-changed-4715eda5-a164-4e5c-9a61-c2317e4046b6. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 527.329485] env[62199]: DEBUG oslo_concurrency.lockutils [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] Acquiring lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.329485] env[62199]: DEBUG oslo_concurrency.lockutils [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] Acquired lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.329724] env[62199]: DEBUG nova.network.neutron [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Refreshing network info cache for port 4715eda5-a164-4e5c-9a61-c2317e4046b6 {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 527.329724] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] Expecting reply to msg c8da37ced4264013ad4f86c8680b40e7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 527.354379] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c8da37ced4264013ad4f86c8680b40e7 [ 527.372319] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8c4a41c85f1b4d38add04ace7de9fb6e [ 527.391016] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c527ffb8f097455ab4444c6259e3ec20 [ 527.526000] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Acquiring lock "11e98624-f84b-420e-b5e1-42dbdd4779e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.526186] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Lock "11e98624-f84b-420e-b5e1-42dbdd4779e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.526694] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg ca702ad5116f4054b58b4471865122eb in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 527.546398] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ca702ad5116f4054b58b4471865122eb [ 527.588512] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398546, 'name': CreateVM_Task} progress is 99%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 527.813736] env[62199]: DEBUG nova.compute.utils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.813736] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg e8f5ffc978ed4c61b9a9c09dd719f3e5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 527.815769] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg bfb652e883dd40db92db615c302ed227 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 527.817499] env[62199]: DEBUG nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Not allocating networking since 'none' was specified. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 527.824215] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bfb652e883dd40db92db615c302ed227 [ 527.840915] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e8f5ffc978ed4c61b9a9c09dd719f3e5 [ 527.890163] env[62199]: DEBUG nova.network.neutron [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 528.028920] env[62199]: DEBUG nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 528.030747] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 4811da7c9dc044428a7749ad3d8e37bc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 528.070484] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4811da7c9dc044428a7749ad3d8e37bc [ 528.088332] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398546, 'name': CreateVM_Task, 'duration_secs': 2.171864} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 528.088504] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Created VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 528.089494] env[62199]: DEBUG oslo_vmware.service [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c86dd012-3a33-40e1-a3f9-3ee57cdabd28 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.095215] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.095376] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.096030] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 528.096278] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-73af148f-a2ca-449b-b2b0-7001023db6ae {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.104255] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 528.104255] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]52d97a97-1746-2a9d-1b78-f398f30f1bcf" [ 528.104255] env[62199]: _type = "Task" [ 528.104255] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.118543] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52d97a97-1746-2a9d-1b78-f398f30f1bcf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.253568] env[62199]: DEBUG nova.network.neutron [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.254078] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] Expecting reply to msg a3fd308f7ac44f4cb1f4858c4bb46821 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 528.262801] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a3fd308f7ac44f4cb1f4858c4bb46821 [ 528.319313] env[62199]: DEBUG nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.321382] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg b2991c8b9c464ceda7fca98dbf93560e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 528.374949] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b2991c8b9c464ceda7fca98dbf93560e [ 528.519875] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e85ee808-b74b-4e8e-8257-765ae7f219f4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.525879] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9d7a3b3-e20f-43b0-82a1-45da6a0d6a21 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.569198] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1583efbb-19d7-4b5f-9755-c986aaa9fd06 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.577492] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967e1eac-7ffb-4761-849d-d88dd5d3d797 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.594347] env[62199]: DEBUG nova.compute.provider_tree [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.594850] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg fc18abe69c474e9e82a3ac6c1c528786 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 528.599179] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.608927] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg fc18abe69c474e9e82a3ac6c1c528786 [ 528.614322] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.614393] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Processing image 9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 528.614605] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 528.614736] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 528.615376] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 528.615620] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f8c03253-1320-41b4-95cf-0807d4702cfa {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.624556] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 528.624724] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62199) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 528.625516] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb5973d-2c1b-43f6-bf84-e6cefd0e5b02 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.635653] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bec7e96d-b90a-4576-92db-9a56b180063f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.643152] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 528.643152] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]5268f0e3-0d7a-487e-4ba2-d27e48c6c9c7" [ 528.643152] env[62199]: _type = "Task" [ 528.643152] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 528.651413] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]5268f0e3-0d7a-487e-4ba2-d27e48c6c9c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 528.756537] env[62199]: DEBUG oslo_concurrency.lockutils [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] Releasing lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.756790] env[62199]: DEBUG nova.compute.manager [req-30dce223-515c-4073-8a79-15409e134455 req-b14cefea-cc43-4625-a726-68f85b782b12 service nova] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Received event network-vif-deleted-4715eda5-a164-4e5c-9a61-c2317e4046b6 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 528.829690] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 1d5bdb1557e7484c8aa9d9105fcb79c3 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 528.882945] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1d5bdb1557e7484c8aa9d9105fcb79c3 [ 529.099998] env[62199]: DEBUG nova.scheduler.client.report [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 529.102417] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg eb4d4dba860a49a09305c12613eab4c4 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 529.114684] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg eb4d4dba860a49a09305c12613eab4c4 [ 529.154966] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Preparing fetch location {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 529.155511] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating directory with path [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 529.155858] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d9f004d2-1fa8-41ba-b963-6ada0078a482 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.189840] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Created directory with path [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 529.190268] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Fetch image to [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 529.190554] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Downloading image file data 9fa830b9-8c21-4a7d-b7cb-64c85631deef to [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk on the data store datastore1 {{(pid=62199) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 529.191496] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fbfd59-cd29-40fc-a725-67a9e06c1d34 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.200178] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5222bd-8e98-4133-a066-299588765cba {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.211019] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73154e54-1021-4ac2-a6ec-57ebaf5f3cc0 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.256517] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3024dd9-a46d-4f28-9ce5-3ae9729457c6 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.262839] env[62199]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-eba14c05-ecbd-49f2-b32b-44dd5f661612 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.291960] env[62199]: DEBUG nova.virt.vmwareapi.images [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Downloading image file data 9fa830b9-8c21-4a7d-b7cb-64c85631deef to the data store datastore1 {{(pid=62199) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 529.331522] env[62199]: DEBUG nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.357356] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.357587] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.357771] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.358025] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.358060] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.358548] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.358548] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.358548] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.358678] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.358897] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.359302] env[62199]: DEBUG nova.virt.hardware [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.360090] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484086c3-a45c-42a1-9085-b65cab0778ad {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.368759] env[62199]: DEBUG oslo_vmware.rw_handles [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62199) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 529.373569] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d92afe09-4297-4339-9626-153edab093b7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.457038] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Instance VIF info [] {{(pid=62199) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 529.463239] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Creating folder: Project (f9dd40b4eea14045a3c54f4d3af0bdfd). Parent ref: group-v108209. {{(pid=62199) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 529.464867] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bea57ff3-bf06-46fe-ad7d-c92d36432a44 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.483256] env[62199]: INFO nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Created folder: Project (f9dd40b4eea14045a3c54f4d3af0bdfd) in parent group-v108209. [ 529.483442] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Creating folder: Instances. Parent ref: group-v108213. {{(pid=62199) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 529.483717] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-452da3ef-63e3-4e9c-9caa-61a6f1cdf95b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.500020] env[62199]: INFO nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Created folder: Instances in parent group-v108213. [ 529.500627] env[62199]: DEBUG oslo.service.loopingcall [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 529.501752] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Creating VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 529.502042] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-39916c30-3e8a-4340-b643-c678ae044c63 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.524550] env[62199]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 529.524550] env[62199]: value = "task-398549" [ 529.524550] env[62199]: _type = "Task" [ 529.524550] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 529.535435] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398549, 'name': CreateVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 529.605483] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.299s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.606390] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 529.607820] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 3a442eba114e42a99f51eae875840e66 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 529.610706] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.201s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.612175] env[62199]: INFO nova.compute.claims [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.613697] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 8ca5573ee3664ae0b12816dc98105556 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 529.662584] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3a442eba114e42a99f51eae875840e66 [ 529.663123] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8ca5573ee3664ae0b12816dc98105556 [ 530.035102] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398549, 'name': CreateVM_Task, 'duration_secs': 0.281367} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 530.035387] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Created VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 530.035836] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.036118] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 530.036318] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 530.036560] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8f7eb84-2bc6-4e56-96e6-dab1ab42790f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.042439] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 530.042439] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]529c2ec5-80ef-b36a-2997-0b7454632523" [ 530.042439] env[62199]: _type = "Task" [ 530.042439] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.053343] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]529c2ec5-80ef-b36a-2997-0b7454632523, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.116915] env[62199]: DEBUG nova.compute.utils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 530.117558] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg bc67387761504cba84c4a6fe3ddced35 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 530.125850] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 16cd834937884152bd9edf944eaf752e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 530.136641] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 530.136831] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 530.139315] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bc67387761504cba84c4a6fe3ddced35 [ 530.141145] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 16cd834937884152bd9edf944eaf752e [ 530.184772] env[62199]: DEBUG oslo_vmware.rw_handles [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Completed reading data from the image iterator. {{(pid=62199) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 530.184980] env[62199]: DEBUG oslo_vmware.rw_handles [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62199) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 530.240209] env[62199]: DEBUG nova.virt.vmwareapi.images [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Downloaded image file data 9fa830b9-8c21-4a7d-b7cb-64c85631deef to vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk on the data store datastore1 {{(pid=62199) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 530.241882] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Caching image {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 530.242170] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copying Virtual Disk [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk to [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 530.242772] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4a10f354-bbaf-4b08-a104-9b0f393257e6 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.254474] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 530.254474] env[62199]: value = "task-398550" [ 530.254474] env[62199]: _type = "Task" [ 530.254474] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 530.262620] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398550, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.488096] env[62199]: DEBUG nova.policy [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '243f12bad8cf4027b1cb79b8d2a7ee8e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '71704572f6c04c94a2fbdab3bd46861a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 530.552177] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.552399] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Processing image 9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 530.552651] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 530.625437] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 530.627279] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg c145efde7a9f45389daf3fe61438e1fe in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 530.694928] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c145efde7a9f45389daf3fe61438e1fe [ 530.764382] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398550, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 530.842514] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-944cac34-8348-4c80-b072-b0b1a284af0f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.850459] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbbb946-620f-4c49-8957-87b82a487662 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.890387] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b574ed4-93f2-4c35-8230-5af25bbaacfe {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.905026] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dc34c93-cad9-4e01-a408-81cc33cf6289 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.919523] env[62199]: DEBUG nova.compute.provider_tree [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.920204] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 4e6f385ab03241d386b42cd202232923 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 530.937535] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4e6f385ab03241d386b42cd202232923 [ 531.059544] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Acquiring lock "0840bc67-7390-4977-8004-f7a09358990d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.059839] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Lock "0840bc67-7390-4977-8004-f7a09358990d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.140228] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg e422108604bd4bc5abae2094852d0aee in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 531.190626] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e422108604bd4bc5abae2094852d0aee [ 531.265075] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398550, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.813357} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.265329] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copied Virtual Disk [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk to [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 531.265512] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleting the datastore file [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 531.265753] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8040e7ee-4894-4721-93e4-fa03a1806c16 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.272400] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 531.272400] env[62199]: value = "task-398551" [ 531.272400] env[62199]: _type = "Task" [ 531.272400] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.280553] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398551, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.429734] env[62199]: DEBUG nova.scheduler.client.report [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.432104] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 89165ea9d62243938c6cc9c87910fe42 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 531.447966] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 89165ea9d62243938c6cc9c87910fe42 [ 531.644308] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 531.681581] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.681581] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.681581] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.681894] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.681894] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.681894] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.681894] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.682016] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.682085] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.682258] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.683876] env[62199]: DEBUG nova.virt.hardware [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.683876] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae606c2-8b07-43e5-a790-b06e99076f62 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.692957] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25fe4449-2840-49b4-9fcc-271ce5d0da53 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.781650] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398551, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023535} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 531.781869] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleted the datastore file {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 531.782083] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Moving file from [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c/9fa830b9-8c21-4a7d-b7cb-64c85631deef to [datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef. {{(pid=62199) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 531.782350] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-65823ee6-4f21-436a-8645-1d4c962e990c {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.789671] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 531.789671] env[62199]: value = "task-398552" [ 531.789671] env[62199]: _type = "Task" [ 531.789671] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 531.800571] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398552, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 531.937305] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.937305] env[62199]: DEBUG nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.937305] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg b366b67ebc01495fb7970999cdc43d70 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 531.939817] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.946s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.939817] env[62199]: INFO nova.compute.claims [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 531.941410] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 8da7b0b1031d4b7f9ea5faca19356643 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 532.001057] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8da7b0b1031d4b7f9ea5faca19356643 [ 532.004654] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b366b67ebc01495fb7970999cdc43d70 [ 532.301294] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398552, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.031525} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.301696] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] File moved {{(pid=62199) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 532.303026] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Cleaning up location [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 532.303026] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleting the datastore file [datastore1] vmware_temp/fd857ea4-8119-4d35-b986-737f3a33529c {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 532.303026] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8f5c9d8e-5826-4a9f-a7a6-3f805927a65f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.309717] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 532.309717] env[62199]: value = "task-398553" [ 532.309717] env[62199]: _type = "Task" [ 532.309717] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.321745] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398553, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.352521] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Successfully created port: e7d41690-0430-4423-a444-832d87c671e3 {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 532.444763] env[62199]: DEBUG nova.compute.utils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.445393] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg d0d6eb3d12bb4591bbdb480c7127ff71 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 532.447391] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 498476ffe7054a428668658b83f374ae in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 532.448335] env[62199]: DEBUG nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.448490] env[62199]: DEBUG nova.network.neutron [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.462162] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 498476ffe7054a428668658b83f374ae [ 532.468547] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d0d6eb3d12bb4591bbdb480c7127ff71 [ 532.583727] env[62199]: DEBUG oslo_concurrency.lockutils [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Acquiring lock "7a0241f3-41cc-494a-9a16-bf95b8abfab9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.583948] env[62199]: DEBUG oslo_concurrency.lockutils [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Lock "7a0241f3-41cc-494a-9a16-bf95b8abfab9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.602360] env[62199]: DEBUG nova.policy [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '8c96900a25b741f185b14da57f679991', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad703de388d34a93a2642c6d866e8634', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 532.821462] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398553, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042175} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 532.821723] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleted the datastore file {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 532.823012] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-69537256-7fec-4efa-9c6f-3c736c843065 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.828708] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 532.828708] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]52af1580-485d-74f1-23c7-ba4b3c6b7651" [ 532.828708] env[62199]: _type = "Task" [ 532.828708] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 532.840202] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52af1580-485d-74f1-23c7-ba4b3c6b7651, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 532.949535] env[62199]: DEBUG nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.951348] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 97f08a520e4d4b3b8292c97d088fbbca in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 532.999361] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 97f08a520e4d4b3b8292c97d088fbbca [ 533.161516] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524c6e8e-bf47-45b5-82bc-2296b6c5a744 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.169508] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c87a2c-0b99-4ba4-a084-c16d6824d6b2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.212455] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61d73334-013c-4395-b05c-9656b6c5a6be {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.228042] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6de632b4-7087-4003-813c-d597d765afb7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.246821] env[62199]: DEBUG nova.compute.provider_tree [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.247380] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 363436ab5e284f35bbadbfd86bcf0c47 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 533.255940] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 363436ab5e284f35bbadbfd86bcf0c47 [ 533.340632] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52af1580-485d-74f1-23c7-ba4b3c6b7651, 'name': SearchDatastore_Task, 'duration_secs': 0.008761} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.340963] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.341151] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore1] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 533.341410] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.341583] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 533.341795] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a26b7d62-1a66-4707-8fd9-74bb843e7eaf {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.343724] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b657e43a-57f3-4980-a7b9-bddd5a203193 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.351471] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 533.351471] env[62199]: value = "task-398554" [ 533.351471] env[62199]: _type = "Task" [ 533.351471] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.358055] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 533.358205] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62199) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 533.359615] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85117e18-b0e7-4b24-b312-d7ac9c2a3bc8 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.372535] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398554, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.379776] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 533.379776] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]52f32cfb-2c47-c9a8-af68-49e231ea88e8" [ 533.379776] env[62199]: _type = "Task" [ 533.379776] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.392279] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52f32cfb-2c47-c9a8-af68-49e231ea88e8, 'name': SearchDatastore_Task, 'duration_secs': 0.008669} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.393192] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-02cfb839-8ecc-4de4-873b-0fd740ecde45 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.398874] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 533.398874] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]52f454d2-2512-f4c3-999c-60927c14ce39" [ 533.398874] env[62199]: _type = "Task" [ 533.398874] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.406867] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52f454d2-2512-f4c3-999c-60927c14ce39, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.447805] env[62199]: ERROR nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. [ 533.447805] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.447805] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.447805] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.447805] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.447805] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.447805] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.447805] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.447805] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.447805] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 533.447805] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.447805] env[62199]: ERROR nova.compute.manager raise self.value [ 533.447805] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.447805] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.447805] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.447805] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.448372] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.448372] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.448372] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. [ 533.448372] env[62199]: ERROR nova.compute.manager [ 533.448372] env[62199]: Traceback (most recent call last): [ 533.448372] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.448372] env[62199]: listener.cb(fileno) [ 533.448372] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.448372] env[62199]: result = function(*args, **kwargs) [ 533.448372] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.448372] env[62199]: return func(*args, **kwargs) [ 533.448372] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.448372] env[62199]: raise e [ 533.448372] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.448372] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 533.448372] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.448372] env[62199]: created_port_ids = self._update_ports_for_instance( [ 533.448372] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.448372] env[62199]: with excutils.save_and_reraise_exception(): [ 533.448372] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.448372] env[62199]: self.force_reraise() [ 533.448372] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.448372] env[62199]: raise self.value [ 533.448372] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.448372] env[62199]: updated_port = self._update_port( [ 533.448372] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.448372] env[62199]: _ensure_no_port_binding_failure(port) [ 533.448372] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.448372] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.449193] env[62199]: nova.exception.PortBindingFailed: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. [ 533.449193] env[62199]: Removing descriptor: 15 [ 533.449193] env[62199]: ERROR nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Traceback (most recent call last): [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] yield resources [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self.driver.spawn(context, instance, image_meta, [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.449193] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] vm_ref = self.build_virtual_machine(instance, [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] for vif in network_info: [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return self._sync_wrapper(fn, *args, **kwargs) [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self.wait() [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self[:] = self._gt.wait() [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return self._exit_event.wait() [ 533.449521] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] result = hub.switch() [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return self.greenlet.switch() [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] result = function(*args, **kwargs) [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return func(*args, **kwargs) [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] raise e [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] nwinfo = self.network_api.allocate_for_instance( [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.449905] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] created_port_ids = self._update_ports_for_instance( [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] with excutils.save_and_reraise_exception(): [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self.force_reraise() [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] raise self.value [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] updated_port = self._update_port( [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] _ensure_no_port_binding_failure(port) [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.450317] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] raise exception.PortBindingFailed(port_id=port['id']) [ 533.450632] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] nova.exception.PortBindingFailed: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. [ 533.450632] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] [ 533.450632] env[62199]: INFO nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Terminating instance [ 533.452120] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Acquiring lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.452120] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Acquired lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.452120] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.452120] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 3468b3fb29b0412291ce9146319a3b30 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 533.456460] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 6925c42733ae42ca91111cd835f44fa6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 533.473439] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3468b3fb29b0412291ce9146319a3b30 [ 533.516207] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6925c42733ae42ca91111cd835f44fa6 [ 533.535165] env[62199]: DEBUG nova.compute.manager [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Received event network-changed-da4c2f45-8714-4d2c-8997-38c3c011820b {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 533.535165] env[62199]: DEBUG nova.compute.manager [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Refreshing instance network info cache due to event network-changed-da4c2f45-8714-4d2c-8997-38c3c011820b. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 533.535165] env[62199]: DEBUG oslo_concurrency.lockutils [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] Acquiring lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.704383] env[62199]: DEBUG oslo_concurrency.lockutils [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Acquiring lock "7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.704547] env[62199]: DEBUG oslo_concurrency.lockutils [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Lock "7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.757064] env[62199]: DEBUG nova.scheduler.client.report [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 533.759658] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg e1b13dec661a43a8ac03c2d8de16c3d8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 533.771947] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e1b13dec661a43a8ac03c2d8de16c3d8 [ 533.861442] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398554, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455536} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.861679] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore1] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 533.861881] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Extending root virtual disk to 1048576 {{(pid=62199) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 533.862121] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1ae01614-cec1-474d-be89-bba09da2d709 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.868888] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 533.868888] env[62199]: value = "task-398555" [ 533.868888] env[62199]: _type = "Task" [ 533.868888] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.883301] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398555, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.910242] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52f454d2-2512-f4c3-999c-60927c14ce39, 'name': SearchDatastore_Task, 'duration_secs': 0.00818} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 533.910531] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.910818] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore1] 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8/83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 533.911154] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a404ab4e-b709-41c7-89fb-a311bbeeaa61 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.919029] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 533.919029] env[62199]: value = "task-398556" [ 533.919029] env[62199]: _type = "Task" [ 533.919029] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 533.942235] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398556, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 533.968762] env[62199]: DEBUG nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.998663] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.998905] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.999095] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.999214] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.999350] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.999486] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.999676] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.999879] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.000068] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.000227] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.000390] env[62199]: DEBUG nova.virt.hardware [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.001303] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd08d630-9087-489a-8611-737dca9ffe38 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.010181] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abcf7527-0bd3-4a10-bb8b-4d00ce262bd9 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.025140] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.262258] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.262776] env[62199]: DEBUG nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 534.264569] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg e79900300df942d79c2f549590fd5a13 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 534.265632] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.702s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.267208] env[62199]: INFO nova.compute.claims [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 534.269227] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 7fec438e58f7451383323c3df4d203f1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 534.309831] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e79900300df942d79c2f549590fd5a13 [ 534.315498] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7fec438e58f7451383323c3df4d203f1 [ 534.379183] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398555, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064275} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.379552] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Extended root virtual disk {{(pid=62199) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 534.380654] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e6cf49-4c4c-48e8-97ec-a09251674880 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.412131] env[62199]: DEBUG nova.virt.vmwareapi.volumeops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk or device None with type sparse {{(pid=62199) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 534.412524] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-22131d8a-d144-4974-86cb-b251e05c5bde {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.442850] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398556, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.432183} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.444576] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore1] 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8/83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 534.444882] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Extending root virtual disk to 1048576 {{(pid=62199) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 534.445265] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 534.445265] env[62199]: value = "task-398557" [ 534.445265] env[62199]: _type = "Task" [ 534.445265] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.445620] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb11ec71-aa5a-4e7c-89c2-476525b222d7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.460310] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398557, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.462082] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 534.462082] env[62199]: value = "task-398558" [ 534.462082] env[62199]: _type = "Task" [ 534.462082] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.470736] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398558, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 534.498963] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.499795] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg d4753802b1214d09915ebdd3f1e7c89a in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 534.510930] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d4753802b1214d09915ebdd3f1e7c89a [ 534.726867] env[62199]: DEBUG nova.network.neutron [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Successfully created port: 446591f2-cdd1-4d8e-9486-e52602fe996e {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 534.773274] env[62199]: DEBUG nova.compute.utils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 534.773936] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg d3db6478f4374391abbe6cdf17528548 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 534.774912] env[62199]: DEBUG nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 534.775087] env[62199]: DEBUG nova.network.neutron [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 534.779585] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 68cea7b4451b4bd59a04699d56653088 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 534.793239] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 68cea7b4451b4bd59a04699d56653088 [ 534.800066] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d3db6478f4374391abbe6cdf17528548 [ 534.964570] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398557, 'name': ReconfigVM_Task, 'duration_secs': 0.286198} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.964570] env[62199]: DEBUG nova.virt.vmwareapi.volumeops [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk or device None with type sparse {{(pid=62199) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 534.964570] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e4d08f5a-fd02-4a98-9b47-c9b3a51528d3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.975111] env[62199]: DEBUG nova.policy [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1bf36d61ac78428dbc59a32c738258a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e46d3885fe1344458c9098972dcc8e6c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 534.993977] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398558, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068137} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 534.993977] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 534.993977] env[62199]: value = "task-398559" [ 534.993977] env[62199]: _type = "Task" [ 534.993977] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 534.995196] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Extended root virtual disk {{(pid=62199) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 534.995401] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-443a5b7c-34d8-49eb-9509-9aa1ea82ec6a {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.003890] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Releasing lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.005989] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.005989] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 535.006644] env[62199]: DEBUG oslo_concurrency.lockutils [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] Acquired lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.006644] env[62199]: DEBUG nova.network.neutron [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Refreshing network info cache for port da4c2f45-8714-4d2c-8997-38c3c011820b {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 535.007374] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] Expecting reply to msg 844a96a283f843dd9d090bbf38de9928 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 535.012423] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f790cadc-62e6-47f7-b66b-dd748c2c8b42 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.043087] env[62199]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8/83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8.vmdk or device None with type sparse {{(pid=62199) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 535.044276] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 844a96a283f843dd9d090bbf38de9928 [ 535.051289] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-acf8e74e-0939-453b-a551-7d5d339ce16b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.070358] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398559, 'name': Rename_Task} progress is 14%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.079501] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 535.079501] env[62199]: value = "task-398560" [ 535.079501] env[62199]: _type = "Task" [ 535.079501] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.082542] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdc7d83-a146-42b8-8b65-61371d8d09f7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.115373] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e34b059b-8edf-4a81-9485-358cefd2bde6 could not be found. [ 535.115733] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 535.115979] env[62199]: INFO nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Took 0.11 seconds to destroy the instance on the hypervisor. [ 535.116460] env[62199]: DEBUG oslo.service.loopingcall [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.120350] env[62199]: DEBUG nova.compute.manager [-] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.120470] env[62199]: DEBUG nova.network.neutron [-] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.123734] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398560, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.280051] env[62199]: DEBUG nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 535.280627] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 4a8446a00be3412a80b7117baf78fb9a in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 535.326070] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4a8446a00be3412a80b7117baf78fb9a [ 535.372448] env[62199]: DEBUG nova.network.neutron [-] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.373963] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg cf699a8e8ba84cd1ad45b1e9d2093ab2 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 535.389937] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg cf699a8e8ba84cd1ad45b1e9d2093ab2 [ 535.462976] env[62199]: DEBUG oslo_concurrency.lockutils [None req-177dc621-e73e-4446-a33e-6f58a72c1b85 tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Acquiring lock "a3aa8b80-5a13-4760-9288-25a05ea89442" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 535.463204] env[62199]: DEBUG oslo_concurrency.lockutils [None req-177dc621-e73e-4446-a33e-6f58a72c1b85 tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Lock "a3aa8b80-5a13-4760-9288-25a05ea89442" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.505367] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398559, 'name': Rename_Task, 'duration_secs': 0.134916} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 535.508221] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powering on the VM {{(pid=62199) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 535.508783] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-afd20826-b4a0-479e-9301-3defa0f2b011 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.521996] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87068120-82c7-47aa-bdeb-8b46f58d470e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.533799] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b877033-3d77-419d-a1dc-7bd795eee9ed {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.538902] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 535.538902] env[62199]: value = "task-398561" [ 535.538902] env[62199]: _type = "Task" [ 535.538902] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 535.571028] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72b9a02-1b49-4270-9000-58857551619e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.581993] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398561, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.587084] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc429747-befd-4031-bbba-0c628c400c17 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.602348] env[62199]: DEBUG nova.compute.provider_tree [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.603072] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg aa5a4977e929410c96ff378b685cc1d8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 535.613806] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg aa5a4977e929410c96ff378b685cc1d8 [ 535.614400] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398560, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 535.751853] env[62199]: DEBUG nova.network.neutron [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.788284] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 996506b0505f469cbb9d9f0d0d930df0 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 535.826204] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 996506b0505f469cbb9d9f0d0d930df0 [ 535.889838] env[62199]: DEBUG nova.network.neutron [-] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.890326] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg c0737ea258e64df2985dbd78c9259218 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 535.901240] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c0737ea258e64df2985dbd78c9259218 [ 536.048790] env[62199]: DEBUG oslo_vmware.api [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398561, 'name': PowerOnVM_Task, 'duration_secs': 0.508385} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.049059] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powered on the VM {{(pid=62199) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 536.049244] env[62199]: INFO nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Took 11.27 seconds to spawn the instance on the hypervisor. [ 536.049488] env[62199]: DEBUG nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Checking state {{(pid=62199) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 536.050474] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcdbec83-2d48-40a0-91a7-c249ac2ddca3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.058766] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg f59b05f44d8d42019ea5972aebdb14ef in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 536.092702] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f59b05f44d8d42019ea5972aebdb14ef [ 536.106240] env[62199]: DEBUG nova.scheduler.client.report [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.110648] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 7b4867ecbb284650bfde5e485c16d059 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 536.116388] env[62199]: DEBUG nova.network.neutron [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Successfully created port: 3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 536.120951] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398560, 'name': ReconfigVM_Task, 'duration_secs': 0.651098} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 536.123028] env[62199]: DEBUG nova.virt.vmwareapi.volumeops [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Reconfigured VM instance instance-00000005 to attach disk [datastore1] 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8/83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8.vmdk or device None with type sparse {{(pid=62199) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 536.123622] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-45a538ff-99a2-446b-aa86-1245dd09138f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.127814] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7b4867ecbb284650bfde5e485c16d059 [ 536.130695] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 536.130695] env[62199]: value = "task-398562" [ 536.130695] env[62199]: _type = "Task" [ 536.130695] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.139929] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398562, 'name': Rename_Task} progress is 5%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.190338] env[62199]: DEBUG nova.network.neutron [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.190874] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] Expecting reply to msg da1e5dd958a34fca9d28a80692dd30a1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 536.199627] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg da1e5dd958a34fca9d28a80692dd30a1 [ 536.291975] env[62199]: DEBUG nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 536.316310] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.316310] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.316310] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.316310] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.316764] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.316764] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.316923] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.317138] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.317525] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.317631] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.317860] env[62199]: DEBUG nova.virt.hardware [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.318792] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b249b471-0a1b-450b-a49a-3921d691d6f5 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.327858] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2e8a738-9195-4dfb-baae-493fd9661731 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.392774] env[62199]: INFO nova.compute.manager [-] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Took 1.27 seconds to deallocate network for instance. [ 536.395558] env[62199]: DEBUG nova.compute.claims [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 536.395558] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.570157] env[62199]: INFO nova.compute.manager [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Took 20.03 seconds to build instance. [ 536.570496] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg ce0bd8c3d81c4a039f159c5896d06d1c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 536.592964] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ce0bd8c3d81c4a039f159c5896d06d1c [ 536.613437] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.613517] env[62199]: DEBUG nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 536.615187] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 7f6c09bfab594594aee0dbdb7d46a19e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 536.618925] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 10.763s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.618925] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg d53420728a43442782e9d9bdb1b264ff in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 536.639976] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398562, 'name': Rename_Task} progress is 14%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.668500] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7f6c09bfab594594aee0dbdb7d46a19e [ 536.693138] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d53420728a43442782e9d9bdb1b264ff [ 536.693833] env[62199]: DEBUG oslo_concurrency.lockutils [req-f8acfdf6-f74f-4b66-bda3-522d0e4675d7 req-b0c9d259-a1a0-4882-bb0d-a10694886030 service nova] Releasing lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.073215] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9021f5bd-9343-4cb4-be95-56ad682d191a tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Lock "47b420a2-5eeb-4f14-9137-fb591a9dcbe6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.542s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.073850] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 49f2fe6750e44a7ea7ec6b4a0951927d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 537.093975] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 49f2fe6750e44a7ea7ec6b4a0951927d [ 537.118217] env[62199]: DEBUG nova.compute.utils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.118797] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 0fa207d9906a4bbcba5ba52e5e43b7b5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 537.119713] env[62199]: DEBUG nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 537.119928] env[62199]: DEBUG nova.network.neutron [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 537.131017] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0fa207d9906a4bbcba5ba52e5e43b7b5 [ 537.145143] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398562, 'name': Rename_Task, 'duration_secs': 0.665594} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.150260] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Powering on the VM {{(pid=62199) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 537.152069] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a22a0915-275d-47f7-8a41-faa1f0d5fbae {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.159114] env[62199]: DEBUG oslo_concurrency.lockutils [None req-4eb1ac2e-b373-4132-9b6d-09898ff087ed tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Acquiring lock "4ddf6b0c-f752-4080-bdef-70d954d2c4f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.159366] env[62199]: DEBUG oslo_concurrency.lockutils [None req-4eb1ac2e-b373-4132-9b6d-09898ff087ed tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Lock "4ddf6b0c-f752-4080-bdef-70d954d2c4f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.166263] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 537.166263] env[62199]: value = "task-398563" [ 537.166263] env[62199]: _type = "Task" [ 537.166263] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.174757] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398563, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.303478] env[62199]: DEBUG nova.compute.manager [req-28a39e68-fa41-495c-b4c7-ea4283397aff req-bef46b5c-4d34-4cae-a2d1-5aa98461d43e service nova] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Received event network-vif-deleted-da4c2f45-8714-4d2c-8997-38c3c011820b {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 537.383322] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42fe3c9d-47f6-423b-9b95-fcf3e8266aaf {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.391201] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4033a070-a5bf-4ffe-957f-12a793376ec2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.422991] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbe45db9-0f70-46e8-9937-b9257f6f4245 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.430325] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c05d071c-16c3-4c59-bfe6-c8111893a326 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.443952] env[62199]: DEBUG nova.compute.provider_tree [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 537.444521] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 0b43c841a57d40618e47f6ff579950c3 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 537.452526] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0b43c841a57d40618e47f6ff579950c3 [ 537.576318] env[62199]: DEBUG nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.577990] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 7c0134988f6744219260b69d449157d1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 537.623893] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7c0134988f6744219260b69d449157d1 [ 537.625334] env[62199]: DEBUG nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 537.626982] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 33d4f27ac3734c28824cd617ecc06170 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 537.629478] env[62199]: DEBUG nova.policy [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1bf36d61ac78428dbc59a32c738258a8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e46d3885fe1344458c9098972dcc8e6c', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 537.675678] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398563, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.718592] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 33d4f27ac3734c28824cd617ecc06170 [ 537.947543] env[62199]: DEBUG nova.scheduler.client.report [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.950186] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 8e33585a818e46c58fc761653f09139d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 537.962586] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8e33585a818e46c58fc761653f09139d [ 538.099204] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.135135] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg a18182f0ab734047ad5c0d92654dff1f in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 538.174600] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a18182f0ab734047ad5c0d92654dff1f [ 538.181702] env[62199]: DEBUG oslo_vmware.api [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398563, 'name': PowerOnVM_Task, 'duration_secs': 0.604935} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 538.181702] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Powered on the VM {{(pid=62199) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 538.181702] env[62199]: INFO nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Took 8.85 seconds to spawn the instance on the hypervisor. [ 538.181702] env[62199]: DEBUG nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Checking state {{(pid=62199) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 538.181702] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac148ace-155e-48aa-8def-b636f80c2783 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.190498] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 2448155d5a2d47fdab0ba411482f4bbf in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 538.254610] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2448155d5a2d47fdab0ba411482f4bbf [ 538.399659] env[62199]: ERROR nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. [ 538.399659] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 538.399659] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.399659] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 538.399659] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.399659] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 538.399659] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.399659] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 538.399659] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.399659] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 538.399659] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.399659] env[62199]: ERROR nova.compute.manager raise self.value [ 538.399659] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.399659] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 538.399659] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.399659] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 538.400220] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.400220] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 538.400220] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. [ 538.400220] env[62199]: ERROR nova.compute.manager [ 538.400220] env[62199]: Traceback (most recent call last): [ 538.400220] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 538.400220] env[62199]: listener.cb(fileno) [ 538.400220] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.400220] env[62199]: result = function(*args, **kwargs) [ 538.400220] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.400220] env[62199]: return func(*args, **kwargs) [ 538.400220] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.400220] env[62199]: raise e [ 538.400220] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.400220] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 538.400220] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.400220] env[62199]: created_port_ids = self._update_ports_for_instance( [ 538.400220] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.400220] env[62199]: with excutils.save_and_reraise_exception(): [ 538.400220] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.400220] env[62199]: self.force_reraise() [ 538.400220] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.400220] env[62199]: raise self.value [ 538.400220] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.400220] env[62199]: updated_port = self._update_port( [ 538.400220] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.400220] env[62199]: _ensure_no_port_binding_failure(port) [ 538.400220] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.400220] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 538.401033] env[62199]: nova.exception.PortBindingFailed: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. [ 538.401033] env[62199]: Removing descriptor: 16 [ 538.401033] env[62199]: ERROR nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Traceback (most recent call last): [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] yield resources [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self.driver.spawn(context, instance, image_meta, [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.401033] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] vm_ref = self.build_virtual_machine(instance, [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] for vif in network_info: [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return self._sync_wrapper(fn, *args, **kwargs) [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self.wait() [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self[:] = self._gt.wait() [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return self._exit_event.wait() [ 538.401449] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] result = hub.switch() [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return self.greenlet.switch() [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] result = function(*args, **kwargs) [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return func(*args, **kwargs) [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] raise e [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] nwinfo = self.network_api.allocate_for_instance( [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.401864] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] created_port_ids = self._update_ports_for_instance( [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] with excutils.save_and_reraise_exception(): [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self.force_reraise() [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] raise self.value [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] updated_port = self._update_port( [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] _ensure_no_port_binding_failure(port) [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.402214] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] raise exception.PortBindingFailed(port_id=port['id']) [ 538.402538] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] nova.exception.PortBindingFailed: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. [ 538.402538] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] [ 538.402538] env[62199]: INFO nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Terminating instance [ 538.403431] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Acquiring lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.403583] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Acquired lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.403747] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 538.404203] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 1a3d58b8ff1e4632af1f8e67ff6efcdc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 538.415427] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1a3d58b8ff1e4632af1f8e67ff6efcdc [ 538.453443] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.836s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.453443] env[62199]: ERROR nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. [ 538.453443] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Traceback (most recent call last): [ 538.453443] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 538.453443] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self.driver.spawn(context, instance, image_meta, [ 538.453443] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 538.453443] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 538.453443] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 538.453443] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] vm_ref = self.build_virtual_machine(instance, [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] for vif in network_info: [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return self._sync_wrapper(fn, *args, **kwargs) [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self.wait() [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self[:] = self._gt.wait() [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return self._exit_event.wait() [ 538.454126] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] result = hub.switch() [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return self.greenlet.switch() [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] result = function(*args, **kwargs) [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] return func(*args, **kwargs) [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] raise e [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] nwinfo = self.network_api.allocate_for_instance( [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 538.454464] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] created_port_ids = self._update_ports_for_instance( [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] with excutils.save_and_reraise_exception(): [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] self.force_reraise() [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] raise self.value [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] updated_port = self._update_port( [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] _ensure_no_port_binding_failure(port) [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 538.454794] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] raise exception.PortBindingFailed(port_id=port['id']) [ 538.455085] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] nova.exception.PortBindingFailed: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. [ 538.455085] env[62199]: ERROR nova.compute.manager [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] [ 538.455085] env[62199]: DEBUG nova.compute.utils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. {{(pid=62199) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 538.455754] env[62199]: DEBUG oslo_concurrency.lockutils [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 11.559s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.456171] env[62199]: DEBUG oslo_concurrency.lockutils [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.456171] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62199) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 538.456356] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.505s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.459031] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 01e0a0e692064c1f949c4242f49abbd9 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 538.464029] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e566e99-ab20-475f-a6ac-e73e4f54f671 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.465943] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Build of instance 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3 was re-scheduled: Binding failed for port 9a060963-bba6-4aa8-968a-e6bef71637e0, please check neutron logs for more information. {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 538.466416] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Unplugging VIFs for instance {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 538.466637] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Acquiring lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.466941] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Acquired lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.467163] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 538.467606] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 239066c23bfe42769d479ba5ed14a956 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 538.480299] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 239066c23bfe42769d479ba5ed14a956 [ 538.483877] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0fc0224-6237-4c45-8c51-e82479ae8ee3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.518107] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-505bab56-8c32-4eda-83e9-34e0d32401da {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.521434] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 01e0a0e692064c1f949c4242f49abbd9 [ 538.526974] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f31ad258-2800-43fc-ad0e-3f0054da82eb {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.564397] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181747MB free_disk=126GB free_vcpus=48 pci_devices=None {{(pid=62199) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 538.564552] env[62199]: DEBUG oslo_concurrency.lockutils [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.565589] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 538.641014] env[62199]: DEBUG nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 538.673043] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 538.673265] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 538.673413] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 538.673585] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 538.673722] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 538.673862] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 538.674060] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 538.674209] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 538.674364] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 538.674526] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 538.674688] env[62199]: DEBUG nova.virt.hardware [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 538.675720] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c99cd54-2474-4321-8f93-cd975adbb189 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.686531] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3638dc5a-3072-4d99-8ae4-8bf801537f16 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.711442] env[62199]: INFO nova.compute.manager [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Took 20.82 seconds to build instance. [ 538.712049] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg bc799d6980ee432a89fb1edfda7ae9c9 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 538.725275] env[62199]: DEBUG nova.network.neutron [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Successfully created port: d17af2cc-d944-48ca-8570-36cc3677dc44 {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 538.735676] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bc799d6980ee432a89fb1edfda7ae9c9 [ 538.800736] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.801558] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 073e3852ead943938c5e6378a8db8909 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 538.810244] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 073e3852ead943938c5e6378a8db8909 [ 538.997972] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2a9ce696-55bf-4faa-aebb-264fe3528c6d tempest-FloatingIPsAssociationNegativeTestJSON-1496629650 tempest-FloatingIPsAssociationNegativeTestJSON-1496629650-project-member] Acquiring lock "4b41ffc1-acc3-4568-b34e-548a0880a30d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.997972] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2a9ce696-55bf-4faa-aebb-264fe3528c6d tempest-FloatingIPsAssociationNegativeTestJSON-1496629650 tempest-FloatingIPsAssociationNegativeTestJSON-1496629650-project-member] Lock "4b41ffc1-acc3-4568-b34e-548a0880a30d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.017891] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.215233] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d9f10a9a-c439-4421-8c2c-8419ea6fbaba tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.331s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.215233] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 4d40dcb40eae4596a26531d4d8a4eeaf in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.248539] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4d40dcb40eae4596a26531d4d8a4eeaf [ 539.304450] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Releasing lock "refresh_cache-0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.304450] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 539.304764] env[62199]: DEBUG nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 539.304764] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 539.321403] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e771c7b-ae50-4bcd-a04c-1080372b7f0b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.327406] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3de34225-9efb-49ad-acc0-1a2d18e92b97 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.361859] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4415f8b-444d-4fb5-a2da-15b6eb02b16b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.369070] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 539.369632] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 556a978b8377421fba42578b038b8c07 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.371951] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b4821d0-1276-4038-89ef-7920b6e1fde4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.379430] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 556a978b8377421fba42578b038b8c07 [ 539.380108] env[62199]: DEBUG nova.network.neutron [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.380623] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 0399dbdcdcba4164b64c1fd978faa801 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.391989] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0399dbdcdcba4164b64c1fd978faa801 [ 539.392644] env[62199]: DEBUG nova.compute.provider_tree [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.393140] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 30e08b61bc704236a871b99d43e54d92 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.407919] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 30e08b61bc704236a871b99d43e54d92 [ 539.495734] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.495734] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg a105530b76664cc998076e6075f829ce in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.505382] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a105530b76664cc998076e6075f829ce [ 539.722786] env[62199]: DEBUG nova.compute.manager [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 539.722786] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 0ba028de8a504741aa616d42796cdf84 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.733855] env[62199]: DEBUG oslo_concurrency.lockutils [None req-86db530f-7859-466c-b178-9758f33340ea tempest-ServerShowV247Test-2074241572 tempest-ServerShowV247Test-2074241572-project-member] Acquiring lock "b56237fc-838d-4ce2-b435-60b4f611f113" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.733855] env[62199]: DEBUG oslo_concurrency.lockutils [None req-86db530f-7859-466c-b178-9758f33340ea tempest-ServerShowV247Test-2074241572 tempest-ServerShowV247Test-2074241572-project-member] Lock "b56237fc-838d-4ce2-b435-60b4f611f113" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.772559] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0ba028de8a504741aa616d42796cdf84 [ 539.884321] env[62199]: INFO nova.compute.manager [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] [instance: 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3] Took 0.58 seconds to deallocate network for instance. [ 539.886043] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg ca27e123271f434e83dcda21bf83ae36 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.898275] env[62199]: DEBUG nova.scheduler.client.report [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.900623] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg e096eb64e041464f8ebfefbb490222ae in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 539.910979] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e096eb64e041464f8ebfefbb490222ae [ 539.929813] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ca27e123271f434e83dcda21bf83ae36 [ 539.997477] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Releasing lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 539.998173] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 539.998413] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 539.998712] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db07d6f5-c97f-4f7e-a5fa-f26a4b04bed3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.007692] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49846ed0-0e8f-4763-bc19-0f74a0dde2b8 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.031924] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f could not be found. [ 540.032248] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 540.032478] env[62199]: INFO nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 540.032919] env[62199]: DEBUG oslo.service.loopingcall [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.033167] env[62199]: DEBUG nova.compute.manager [-] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.033314] env[62199]: DEBUG nova.network.neutron [-] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 540.076531] env[62199]: DEBUG nova.network.neutron [-] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.077103] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 037c495e6c814123abb22c41626d1f54 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.085658] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 037c495e6c814123abb22c41626d1f54 [ 540.152803] env[62199]: ERROR nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 446591f2-cdd1-4d8e-9486-e52602fe996e, please check neutron logs for more information. [ 540.152803] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.152803] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.152803] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.152803] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.152803] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.152803] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.152803] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.152803] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.152803] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 540.152803] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.152803] env[62199]: ERROR nova.compute.manager raise self.value [ 540.152803] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.152803] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.152803] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.152803] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.153222] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.153222] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.153222] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 446591f2-cdd1-4d8e-9486-e52602fe996e, please check neutron logs for more information. [ 540.153222] env[62199]: ERROR nova.compute.manager [ 540.153222] env[62199]: Traceback (most recent call last): [ 540.153222] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.153222] env[62199]: listener.cb(fileno) [ 540.153222] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.153222] env[62199]: result = function(*args, **kwargs) [ 540.153222] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.153222] env[62199]: return func(*args, **kwargs) [ 540.153222] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.153222] env[62199]: raise e [ 540.153222] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.153222] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 540.153222] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.153222] env[62199]: created_port_ids = self._update_ports_for_instance( [ 540.153222] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.153222] env[62199]: with excutils.save_and_reraise_exception(): [ 540.153222] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.153222] env[62199]: self.force_reraise() [ 540.153222] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.153222] env[62199]: raise self.value [ 540.153222] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.153222] env[62199]: updated_port = self._update_port( [ 540.153222] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.153222] env[62199]: _ensure_no_port_binding_failure(port) [ 540.153222] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.153222] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.153930] env[62199]: nova.exception.PortBindingFailed: Binding failed for port 446591f2-cdd1-4d8e-9486-e52602fe996e, please check neutron logs for more information. [ 540.153930] env[62199]: Removing descriptor: 18 [ 540.153930] env[62199]: ERROR nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 446591f2-cdd1-4d8e-9486-e52602fe996e, please check neutron logs for more information. [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Traceback (most recent call last): [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] yield resources [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] self.driver.spawn(context, instance, image_meta, [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.153930] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] vm_ref = self.build_virtual_machine(instance, [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] for vif in network_info: [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] return self._sync_wrapper(fn, *args, **kwargs) [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] self.wait() [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] self[:] = self._gt.wait() [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] return self._exit_event.wait() [ 540.154287] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] result = hub.switch() [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] return self.greenlet.switch() [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] result = function(*args, **kwargs) [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] return func(*args, **kwargs) [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] raise e [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] nwinfo = self.network_api.allocate_for_instance( [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.154660] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] created_port_ids = self._update_ports_for_instance( [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] with excutils.save_and_reraise_exception(): [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] self.force_reraise() [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] raise self.value [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] updated_port = self._update_port( [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] _ensure_no_port_binding_failure(port) [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.154990] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] raise exception.PortBindingFailed(port_id=port['id']) [ 540.155362] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] nova.exception.PortBindingFailed: Binding failed for port 446591f2-cdd1-4d8e-9486-e52602fe996e, please check neutron logs for more information. [ 540.155362] env[62199]: ERROR nova.compute.manager [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] [ 540.155362] env[62199]: INFO nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Terminating instance [ 540.155362] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Acquiring lock "refresh_cache-52bd7ff6-1149-4a56-95c2-806572b526f4" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.155362] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Acquired lock "refresh_cache-52bd7ff6-1149-4a56-95c2-806572b526f4" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.155362] env[62199]: DEBUG nova.network.neutron [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.155534] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 42cb2af27b87414fb9d8e335dc332373 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.159947] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 42cb2af27b87414fb9d8e335dc332373 [ 540.244480] env[62199]: DEBUG oslo_concurrency.lockutils [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.390790] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 5b81fb76671b49c2a0cb0a0c9ca0f96e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.408036] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.951s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.412196] env[62199]: ERROR nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Traceback (most recent call last): [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self.driver.spawn(context, instance, image_meta, [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] vm_ref = self.build_virtual_machine(instance, [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.412196] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] for vif in network_info: [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return self._sync_wrapper(fn, *args, **kwargs) [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self.wait() [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self[:] = self._gt.wait() [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return self._exit_event.wait() [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] result = hub.switch() [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.412628] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return self.greenlet.switch() [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] result = function(*args, **kwargs) [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] return func(*args, **kwargs) [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] raise e [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] nwinfo = self.network_api.allocate_for_instance( [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] created_port_ids = self._update_ports_for_instance( [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] with excutils.save_and_reraise_exception(): [ 540.412963] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] self.force_reraise() [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] raise self.value [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] updated_port = self._update_port( [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] _ensure_no_port_binding_failure(port) [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] raise exception.PortBindingFailed(port_id=port['id']) [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] nova.exception.PortBindingFailed: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. [ 540.413274] env[62199]: ERROR nova.compute.manager [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] [ 540.413540] env[62199]: DEBUG nova.compute.utils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. {{(pid=62199) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 540.413540] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.812s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.413540] env[62199]: INFO nova.compute.claims [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.415404] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 5212aa6dda3f42fbacaece2a9a540928 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.417316] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Build of instance 7384f309-7b32-40fc-881e-b967c7b8fec6 was re-scheduled: Binding failed for port 4715eda5-a164-4e5c-9a61-c2317e4046b6, please check neutron logs for more information. {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 540.418025] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Unplugging VIFs for instance {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 540.418395] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Acquiring lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.418692] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Acquired lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.418995] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.419636] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg c39ad9be23a64a6a89d393f8dc3f89b8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.440718] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c39ad9be23a64a6a89d393f8dc3f89b8 [ 540.443294] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5b81fb76671b49c2a0cb0a0c9ca0f96e [ 540.478921] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5212aa6dda3f42fbacaece2a9a540928 [ 540.556434] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg c55b3afaa76f4181a6c84470edc276ff in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.570532] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c55b3afaa76f4181a6c84470edc276ff [ 540.579098] env[62199]: DEBUG nova.network.neutron [-] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.579608] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 5802b87fcef643c0bb8bc1eba0b64dca in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.593821] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5802b87fcef643c0bb8bc1eba0b64dca [ 540.615159] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d1170f4d-991c-464d-9b98-e1c9b40cda00 tempest-ServerDiskConfigTestJSON-1641201365 tempest-ServerDiskConfigTestJSON-1641201365-project-member] Acquiring lock "ccaed194-e8b4-4620-be97-588a232af696" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.615390] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d1170f4d-991c-464d-9b98-e1c9b40cda00 tempest-ServerDiskConfigTestJSON-1641201365 tempest-ServerDiskConfigTestJSON-1641201365-project-member] Lock "ccaed194-e8b4-4620-be97-588a232af696" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.771721] env[62199]: DEBUG nova.compute.manager [None req-1704895a-eaed-4a36-b103-24588ba6bc7e tempest-ServerDiagnosticsV248Test-1307978044 tempest-ServerDiagnosticsV248Test-1307978044-project-admin] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Checking state {{(pid=62199) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 540.771721] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afd2e390-80e5-40c9-9386-ee78e10d0b45 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.775400] env[62199]: INFO nova.compute.manager [None req-1704895a-eaed-4a36-b103-24588ba6bc7e tempest-ServerDiagnosticsV248Test-1307978044 tempest-ServerDiagnosticsV248Test-1307978044-project-admin] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Retrieving diagnostics [ 540.776660] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af303e81-f111-4d42-b148-dad9d847f8f4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.816926] env[62199]: DEBUG nova.network.neutron [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.826702] env[62199]: ERROR nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f, please check neutron logs for more information. [ 540.826702] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 540.826702] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.826702] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 540.826702] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.826702] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 540.826702] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.826702] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 540.826702] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.826702] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 540.826702] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.826702] env[62199]: ERROR nova.compute.manager raise self.value [ 540.826702] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.826702] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 540.826702] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.826702] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 540.827140] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.827140] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 540.827140] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f, please check neutron logs for more information. [ 540.827140] env[62199]: ERROR nova.compute.manager [ 540.827140] env[62199]: Traceback (most recent call last): [ 540.827140] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 540.827140] env[62199]: listener.cb(fileno) [ 540.827140] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.827140] env[62199]: result = function(*args, **kwargs) [ 540.827140] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.827140] env[62199]: return func(*args, **kwargs) [ 540.827140] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.827140] env[62199]: raise e [ 540.827140] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.827140] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 540.827140] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.827140] env[62199]: created_port_ids = self._update_ports_for_instance( [ 540.827140] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.827140] env[62199]: with excutils.save_and_reraise_exception(): [ 540.827140] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.827140] env[62199]: self.force_reraise() [ 540.827140] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.827140] env[62199]: raise self.value [ 540.827140] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.827140] env[62199]: updated_port = self._update_port( [ 540.827140] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.827140] env[62199]: _ensure_no_port_binding_failure(port) [ 540.827140] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.827140] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 540.827843] env[62199]: nova.exception.PortBindingFailed: Binding failed for port 3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f, please check neutron logs for more information. [ 540.827843] env[62199]: Removing descriptor: 15 [ 540.827843] env[62199]: ERROR nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f, please check neutron logs for more information. [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Traceback (most recent call last): [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] yield resources [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] self.driver.spawn(context, instance, image_meta, [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.827843] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] vm_ref = self.build_virtual_machine(instance, [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] for vif in network_info: [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] return self._sync_wrapper(fn, *args, **kwargs) [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] self.wait() [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] self[:] = self._gt.wait() [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] return self._exit_event.wait() [ 540.828221] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] result = hub.switch() [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] return self.greenlet.switch() [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] result = function(*args, **kwargs) [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] return func(*args, **kwargs) [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] raise e [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] nwinfo = self.network_api.allocate_for_instance( [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.828659] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] created_port_ids = self._update_ports_for_instance( [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] with excutils.save_and_reraise_exception(): [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] self.force_reraise() [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] raise self.value [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] updated_port = self._update_port( [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] _ensure_no_port_binding_failure(port) [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.829045] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] raise exception.PortBindingFailed(port_id=port['id']) [ 540.829459] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] nova.exception.PortBindingFailed: Binding failed for port 3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f, please check neutron logs for more information. [ 540.829459] env[62199]: ERROR nova.compute.manager [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] [ 540.829459] env[62199]: INFO nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Terminating instance [ 540.836344] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "refresh_cache-a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.836344] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquired lock "refresh_cache-a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.836344] env[62199]: DEBUG nova.network.neutron [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.836344] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 0f2bcf3ea5ce46a39d7203e129e66358 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.844159] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0f2bcf3ea5ce46a39d7203e129e66358 [ 540.917982] env[62199]: INFO nova.scheduler.client.report [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Deleted allocations for instance 0fb7b132-2c4f-43cd-a5b2-e542e3e748e3 [ 540.924082] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 985efa474cbb4b12980cb57871bc1460 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.927158] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Expecting reply to msg 438bac1beb5441a297f1e0ba7f8fd990 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.930693] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 985efa474cbb4b12980cb57871bc1460 [ 540.953568] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 540.957953] env[62199]: DEBUG nova.network.neutron [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.958633] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg c772c5cf0d2b48d49eaa1473f7a12e20 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 540.967304] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c772c5cf0d2b48d49eaa1473f7a12e20 [ 540.975385] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 438bac1beb5441a297f1e0ba7f8fd990 [ 541.059634] env[62199]: INFO nova.compute.manager [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Rebuilding instance [ 541.082500] env[62199]: INFO nova.compute.manager [-] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Took 1.05 seconds to deallocate network for instance. [ 541.085357] env[62199]: DEBUG nova.compute.claims [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 541.085573] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.103065] env[62199]: DEBUG nova.compute.manager [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Checking state {{(pid=62199) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 541.104343] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6614f351-aa55-4ed2-ba12-bd29007bd490 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.114414] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg a78192b970fc403ea231fd2eb4d82fd5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.166308] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a78192b970fc403ea231fd2eb4d82fd5 [ 541.178092] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.178092] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 3d54b5f19a05410d9f85a52d317ea4b2 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.194788] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3d54b5f19a05410d9f85a52d317ea4b2 [ 541.289951] env[62199]: DEBUG nova.compute.manager [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Received event network-changed-e7d41690-0430-4423-a444-832d87c671e3 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 541.290540] env[62199]: DEBUG nova.compute.manager [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Refreshing instance network info cache due to event network-changed-e7d41690-0430-4423-a444-832d87c671e3. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 541.290540] env[62199]: DEBUG oslo_concurrency.lockutils [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] Acquiring lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 541.290540] env[62199]: DEBUG oslo_concurrency.lockutils [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] Acquired lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 541.290652] env[62199]: DEBUG nova.network.neutron [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Refreshing network info cache for port e7d41690-0430-4423-a444-832d87c671e3 {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 541.291409] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] Expecting reply to msg 0eaab208611b4f7da6fc05a26185a2af in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.302474] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0eaab208611b4f7da6fc05a26185a2af [ 541.392063] env[62199]: DEBUG nova.network.neutron [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.430372] env[62199]: DEBUG oslo_concurrency.lockutils [None req-6f920d4a-cb74-44be-bb5f-1d2d5de44381 tempest-ServerDiagnosticsTest-2061445243 tempest-ServerDiagnosticsTest-2061445243-project-member] Lock "0fb7b132-2c4f-43cd-a5b2-e542e3e748e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.474s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.430913] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg 7adab7b6d7524f509f61bc8d99a9e531 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.444115] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7adab7b6d7524f509f61bc8d99a9e531 [ 541.463548] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Releasing lock "refresh_cache-52bd7ff6-1149-4a56-95c2-806572b526f4" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.463969] env[62199]: DEBUG nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 541.464177] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 541.464757] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b9e82f6d-b6cb-490a-8156-7e96a8808044 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.480059] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9c852f-58fd-489b-9028-eee76bdbcff0 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.514606] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 52bd7ff6-1149-4a56-95c2-806572b526f4 could not be found. [ 541.514988] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 541.515320] env[62199]: INFO nova.compute.manager [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 541.515699] env[62199]: DEBUG oslo.service.loopingcall [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 541.521145] env[62199]: DEBUG nova.compute.manager [-] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.521367] env[62199]: DEBUG nova.network.neutron [-] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 541.563552] env[62199]: DEBUG nova.network.neutron [-] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.564074] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg b525c3441f1f4f858d556410ca899f1c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.577029] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b525c3441f1f4f858d556410ca899f1c [ 541.618329] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powering off the VM {{(pid=62199) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 541.618787] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f9a0193-4687-489f-a63a-2d4ac048d81b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.632040] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 541.632040] env[62199]: value = "task-398564" [ 541.632040] env[62199]: _type = "Task" [ 541.632040] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.635186] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398564, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.649469] env[62199]: DEBUG nova.network.neutron [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.650614] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg ecd6af4b0e0f486bbe7af5cdce72aa14 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.660236] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ecd6af4b0e0f486bbe7af5cdce72aa14 [ 541.682927] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Releasing lock "refresh_cache-7384f309-7b32-40fc-881e-b967c7b8fec6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.683181] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 541.683381] env[62199]: DEBUG nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 541.683600] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 541.743754] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.744397] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg a18fa4f8ecfa4dfbb7e81e12587fa5ae in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.753486] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a18fa4f8ecfa4dfbb7e81e12587fa5ae [ 541.762415] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb70a742-8202-4269-99ef-bb8875e4404b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.771766] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818ed4ee-4e27-4927-9598-58cf533237ec {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.814696] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20cb1f89-8fa8-4844-bb93-b2f79e806bb0 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.822638] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43187ade-22de-4c82-b4ff-c6686991c82e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.839949] env[62199]: DEBUG nova.compute.provider_tree [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.840687] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 75588b0d2bf44d00b9013799a84fab50 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.849062] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 75588b0d2bf44d00b9013799a84fab50 [ 541.865496] env[62199]: DEBUG nova.network.neutron [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.933825] env[62199]: DEBUG nova.compute.manager [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] [instance: 7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 541.935700] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg a807db7a58b843efaa4afd67ab2fc3c1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 541.975374] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a807db7a58b843efaa4afd67ab2fc3c1 [ 542.022143] env[62199]: ERROR nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d17af2cc-d944-48ca-8570-36cc3677dc44, please check neutron logs for more information. [ 542.022143] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 542.022143] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.022143] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 542.022143] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.022143] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 542.022143] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.022143] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 542.022143] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.022143] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 542.022143] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.022143] env[62199]: ERROR nova.compute.manager raise self.value [ 542.022143] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.022143] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 542.022143] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.022143] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 542.022552] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.022552] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 542.022552] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d17af2cc-d944-48ca-8570-36cc3677dc44, please check neutron logs for more information. [ 542.022552] env[62199]: ERROR nova.compute.manager [ 542.022552] env[62199]: Traceback (most recent call last): [ 542.022552] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 542.022552] env[62199]: listener.cb(fileno) [ 542.022552] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.022552] env[62199]: result = function(*args, **kwargs) [ 542.022552] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.022552] env[62199]: return func(*args, **kwargs) [ 542.022552] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.022552] env[62199]: raise e [ 542.022552] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.022552] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 542.022552] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.022552] env[62199]: created_port_ids = self._update_ports_for_instance( [ 542.022552] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.022552] env[62199]: with excutils.save_and_reraise_exception(): [ 542.022552] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.022552] env[62199]: self.force_reraise() [ 542.022552] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.022552] env[62199]: raise self.value [ 542.022552] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.022552] env[62199]: updated_port = self._update_port( [ 542.022552] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.022552] env[62199]: _ensure_no_port_binding_failure(port) [ 542.022552] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.022552] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 542.023333] env[62199]: nova.exception.PortBindingFailed: Binding failed for port d17af2cc-d944-48ca-8570-36cc3677dc44, please check neutron logs for more information. [ 542.023333] env[62199]: Removing descriptor: 20 [ 542.023333] env[62199]: ERROR nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d17af2cc-d944-48ca-8570-36cc3677dc44, please check neutron logs for more information. [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Traceback (most recent call last): [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] yield resources [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] self.driver.spawn(context, instance, image_meta, [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 542.023333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] vm_ref = self.build_virtual_machine(instance, [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] vif_infos = vmwarevif.get_vif_info(self._session, [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] for vif in network_info: [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] return self._sync_wrapper(fn, *args, **kwargs) [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] self.wait() [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] self[:] = self._gt.wait() [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] return self._exit_event.wait() [ 542.023647] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] result = hub.switch() [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] return self.greenlet.switch() [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] result = function(*args, **kwargs) [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] return func(*args, **kwargs) [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] raise e [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] nwinfo = self.network_api.allocate_for_instance( [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 542.023975] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] created_port_ids = self._update_ports_for_instance( [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] with excutils.save_and_reraise_exception(): [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] self.force_reraise() [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] raise self.value [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] updated_port = self._update_port( [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] _ensure_no_port_binding_failure(port) [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 542.024333] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] raise exception.PortBindingFailed(port_id=port['id']) [ 542.024640] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] nova.exception.PortBindingFailed: Binding failed for port d17af2cc-d944-48ca-8570-36cc3677dc44, please check neutron logs for more information. [ 542.024640] env[62199]: ERROR nova.compute.manager [instance: 2681b401-47b3-4b05-b923-65e3078074d5] [ 542.024640] env[62199]: INFO nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Terminating instance [ 542.025382] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "refresh_cache-2681b401-47b3-4b05-b923-65e3078074d5" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.025536] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquired lock "refresh_cache-2681b401-47b3-4b05-b923-65e3078074d5" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.025699] env[62199]: DEBUG nova.network.neutron [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 542.026100] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 87e6825bd9b34704a6ba4de7d0dd7738 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.032625] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 87e6825bd9b34704a6ba4de7d0dd7738 [ 542.069258] env[62199]: DEBUG nova.network.neutron [-] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.069258] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg e17eeae20e054a6bab96af0cac622cb1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.075255] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e17eeae20e054a6bab96af0cac622cb1 [ 542.137467] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398564, 'name': PowerOffVM_Task, 'duration_secs': 0.119475} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.137998] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powered off the VM {{(pid=62199) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 542.138461] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 542.139315] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-454a1cea-86a7-4281-9f4e-63f896f1833e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.146753] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Unregistering the VM {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 542.147090] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-af7c1fe6-b44f-4d94-887b-d9c0357cc0ae {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.155523] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Releasing lock "refresh_cache-a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.156057] env[62199]: DEBUG nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 542.156465] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 542.160103] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aa24ca51-de97-406b-83bf-a28e021f8be4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.166652] env[62199]: DEBUG nova.network.neutron [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.168905] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] Expecting reply to msg d15df79cc73a4ed78ab318b11d8789d8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.172407] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-536e8122-da21-4111-8476-e793d029166c {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.184965] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d15df79cc73a4ed78ab318b11d8789d8 [ 542.185706] env[62199]: DEBUG oslo_concurrency.lockutils [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] Releasing lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.185922] env[62199]: DEBUG nova.compute.manager [req-3abb1b80-4baa-4603-ae47-d0bcbaa54597 req-121470d3-bd07-469b-8275-f44bf52de7be service nova] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Received event network-vif-deleted-e7d41690-0430-4423-a444-832d87c671e3 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 542.186295] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Unregistered the VM {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 542.186478] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Deleting contents of the VM from datastore datastore1 {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 542.186632] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleting the datastore file [datastore1] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6 {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 542.187469] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-440b1a63-6f92-4032-ba6e-54e052e3cc88 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.198183] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a9cc78e1-979e-4b91-9ad1-31c5c5b342f7 could not be found. [ 542.198386] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.198561] env[62199]: INFO nova.compute.manager [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 542.198796] env[62199]: DEBUG oslo.service.loopingcall [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 542.200029] env[62199]: DEBUG nova.compute.manager [-] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 542.200125] env[62199]: DEBUG nova.network.neutron [-] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 542.201748] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 542.201748] env[62199]: value = "task-398566" [ 542.201748] env[62199]: _type = "Task" [ 542.201748] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.247182] env[62199]: DEBUG nova.network.neutron [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.247735] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg bfef5bc456034440af23c84831508e7b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.258326] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bfef5bc456034440af23c84831508e7b [ 542.309015] env[62199]: DEBUG nova.network.neutron [-] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.309635] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 8bb0dd5bd66347bca41401ab6943d839 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.321329] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8bb0dd5bd66347bca41401ab6943d839 [ 542.348024] env[62199]: DEBUG nova.scheduler.client.report [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.350471] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 7a512d637069425d9861918b0ff1e45b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.372814] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7a512d637069425d9861918b0ff1e45b [ 542.454598] env[62199]: DEBUG oslo_concurrency.lockutils [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.559184] env[62199]: DEBUG nova.network.neutron [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 542.570713] env[62199]: INFO nova.compute.manager [-] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Took 1.05 seconds to deallocate network for instance. [ 542.575962] env[62199]: DEBUG nova.compute.claims [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 542.575962] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.711478] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398566, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.117445} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.711726] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleted the datastore file {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 542.711904] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Deleted contents of the VM from datastore datastore1 {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 542.712094] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 542.715508] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg 3a7ef23fb4444dab8a251e6161b68b79 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.750456] env[62199]: INFO nova.compute.manager [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] [instance: 7384f309-7b32-40fc-881e-b967c7b8fec6] Took 1.07 seconds to deallocate network for instance. [ 542.752165] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 90be75e659494d318b3766971a0895bc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.783828] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3a7ef23fb4444dab8a251e6161b68b79 [ 542.787304] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 90be75e659494d318b3766971a0895bc [ 542.808269] env[62199]: DEBUG nova.network.neutron [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.808269] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Expecting reply to msg 4ff76f039dd2493ba860e82eb350adc6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.816974] env[62199]: DEBUG nova.network.neutron [-] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.817371] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 9dd0041f0a4042e59c21b37fb9484408 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.820063] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4ff76f039dd2493ba860e82eb350adc6 [ 542.836860] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9dd0041f0a4042e59c21b37fb9484408 [ 542.853942] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.854472] env[62199]: DEBUG nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 542.856501] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 307ffd267576431392a0304ee2827920 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.857428] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 6.462s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.859118] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg f0511ab9088d4ce787ce637fb0da2797 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 542.902663] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f0511ab9088d4ce787ce637fb0da2797 [ 542.930223] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 307ffd267576431392a0304ee2827920 [ 543.007511] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9634222d-e1a3-42ec-ad79-2f64faf7ba09 tempest-ServerShowV247Test-2074241572 tempest-ServerShowV247Test-2074241572-project-member] Acquiring lock "e047a038-bcf1-4f90-929a-465bbee3ef67" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.007731] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9634222d-e1a3-42ec-ad79-2f64faf7ba09 tempest-ServerShowV247Test-2074241572 tempest-ServerShowV247Test-2074241572-project-member] Lock "e047a038-bcf1-4f90-929a-465bbee3ef67" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.121884] env[62199]: DEBUG oslo_concurrency.lockutils [None req-7e324abd-949b-421d-a6bd-aa3720dd8c1e tempest-ServersTestJSON-1307205334 tempest-ServersTestJSON-1307205334-project-member] Acquiring lock "77de9475-77f5-4833-aaa4-8a1d6de6db9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.122105] env[62199]: DEBUG oslo_concurrency.lockutils [None req-7e324abd-949b-421d-a6bd-aa3720dd8c1e tempest-ServersTestJSON-1307205334 tempest-ServersTestJSON-1307205334-project-member] Lock "77de9475-77f5-4833-aaa4-8a1d6de6db9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.219419] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg 07ab11abe52e46b78001db2420db3a77 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.257870] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg b81477d35bff4f9ab943ad85e1dc9342 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.279749] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 07ab11abe52e46b78001db2420db3a77 [ 543.302615] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b81477d35bff4f9ab943ad85e1dc9342 [ 543.310124] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Releasing lock "refresh_cache-2681b401-47b3-4b05-b923-65e3078074d5" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.310534] env[62199]: DEBUG nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 543.310714] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 543.311621] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc2659df-e3a2-47dc-9122-a355dd5436ee {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.320517] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702b9d9f-78b5-4897-b047-ba03f242afd4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.331328] env[62199]: INFO nova.compute.manager [-] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Took 1.13 seconds to deallocate network for instance. [ 543.335768] env[62199]: DEBUG nova.compute.claims [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 543.335982] env[62199]: DEBUG oslo_concurrency.lockutils [None req-57f04e16-0edc-4b9e-a6b2-d35e75237937 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 543.346633] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2681b401-47b3-4b05-b923-65e3078074d5 could not be found. [ 543.346883] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 543.347150] env[62199]: INFO nova.compute.manager [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 543.347435] env[62199]: DEBUG oslo.service.loopingcall [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.347710] env[62199]: DEBUG nova.compute.manager [-] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.347840] env[62199]: DEBUG nova.network.neutron [-] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 543.365055] env[62199]: DEBUG nova.compute.utils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.365768] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 6f3be4da4f164a78afaa1a9d8ffa3a6b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.371644] env[62199]: DEBUG nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 543.372274] env[62199]: DEBUG nova.network.neutron [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 543.384680] env[62199]: DEBUG nova.network.neutron [-] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.385188] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg a4b2df5044ef40549641c50fa87f85b1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.386223] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6f3be4da4f164a78afaa1a9d8ffa3a6b [ 543.386968] env[62199]: DEBUG nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 543.388792] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 3be4d5d4a86a47bd80a862f880bb3a43 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.394086] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a4b2df5044ef40549641c50fa87f85b1 [ 543.441907] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3be4d5d4a86a47bd80a862f880bb3a43 [ 543.581244] env[62199]: DEBUG nova.policy [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9f50f331cd534d03909c89c3df567ccd', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '025b1b04cb22429b82726baafaa9e377', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 543.671591] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4507852a-f816-45b8-a1e2-a70bdb0fb772 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.679666] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a97fff4-a3c4-4ecc-8726-0d917c51cf69 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.713615] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efad1c4f-eafc-4e72-8510-e326c9fb6d67 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.721132] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d80ee4e0-80fb-4963-a0cc-6a4c81502565 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.737191] env[62199]: DEBUG nova.compute.provider_tree [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.737742] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg f6802a61cb2242f89be6749b011971e7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.748669] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f6802a61cb2242f89be6749b011971e7 [ 543.753542] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 543.753782] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 543.753919] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 543.754095] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 543.754241] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 543.754386] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 543.754589] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 543.754744] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 543.756563] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 543.756563] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 543.756563] env[62199]: DEBUG nova.virt.hardware [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 543.757587] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64120a1e-2298-4a78-b627-26259435ecb2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.769287] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1459535a-9bdb-4621-9fad-d8cfbe8bd0e2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.784028] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Instance VIF info [] {{(pid=62199) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 543.790037] env[62199]: DEBUG oslo.service.loopingcall [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 543.791136] env[62199]: INFO nova.scheduler.client.report [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Deleted allocations for instance 7384f309-7b32-40fc-881e-b967c7b8fec6 [ 543.796703] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Creating VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 543.796934] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af36c6c3-485d-496e-967b-da80f717d5aa {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.811636] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Expecting reply to msg 49da1cd7c03a437e8dece4c7059c1716 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.819091] env[62199]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 543.819091] env[62199]: value = "task-398567" [ 543.819091] env[62199]: _type = "Task" [ 543.819091] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.824154] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 49da1cd7c03a437e8dece4c7059c1716 [ 543.832093] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398567, 'name': CreateVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.899474] env[62199]: DEBUG nova.network.neutron [-] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.899474] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg c26148b74eff41d5990efeda7afe197e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.905604] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg b066aa8dafe043d8a2232d3b8c16a9fc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 543.921331] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c26148b74eff41d5990efeda7afe197e [ 543.979647] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b066aa8dafe043d8a2232d3b8c16a9fc [ 544.243716] env[62199]: DEBUG nova.scheduler.client.report [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.246104] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 1c08fb53e9e044bfb3cb980d197efaa8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 544.267292] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1c08fb53e9e044bfb3cb980d197efaa8 [ 544.314324] env[62199]: DEBUG oslo_concurrency.lockutils [None req-37039741-b1d3-4efd-917a-087d1d1e38b5 tempest-FloatingIPsAssociationTestJSON-50872365 tempest-FloatingIPsAssociationTestJSON-50872365-project-member] Lock "7384f309-7b32-40fc-881e-b967c7b8fec6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.212s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.314958] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-177dc621-e73e-4446-a33e-6f58a72c1b85 tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Expecting reply to msg 79bc5e98f73846cf95da92b72992c8ef in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 544.329713] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398567, 'name': CreateVM_Task, 'duration_secs': 0.284673} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.330283] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 79bc5e98f73846cf95da92b72992c8ef [ 544.330628] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Created VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 544.331528] env[62199]: DEBUG oslo_vmware.service [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02c151f9-0f00-4678-8f74-a13b77d7e9cd {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.337870] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.338040] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.338426] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 544.338925] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8f86ed58-9a3e-47fe-8f3c-36bf841f3d00 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.343918] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 544.343918] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]527a9f08-fc39-94e3-b4fe-8c003cb2f742" [ 544.343918] env[62199]: _type = "Task" [ 544.343918] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.353743] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]527a9f08-fc39-94e3-b4fe-8c003cb2f742, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.409729] env[62199]: DEBUG nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 544.411017] env[62199]: INFO nova.compute.manager [-] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Took 1.06 seconds to deallocate network for instance. [ 544.413979] env[62199]: DEBUG nova.compute.claims [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 544.413979] env[62199]: DEBUG oslo_concurrency.lockutils [None req-785a2eb6-4be0-4a03-90dc-e65720e74a49 tempest-ServersAdminTestJSON-463259247 tempest-ServersAdminTestJSON-463259247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.438141] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.438141] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.438302] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.438414] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.438557] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.438698] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.438905] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.439056] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.439212] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.439365] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.440387] env[62199]: DEBUG nova.virt.hardware [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.440600] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67774856-5b7a-4921-80c9-ac1f593bcfc3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.448765] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f4e5267-d77b-46fc-8afc-733f2ad8f13d {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.473866] env[62199]: DEBUG nova.network.neutron [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Successfully created port: 28f332db-d316-4659-a57c-3180ea938864 {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 544.750307] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.751005] env[62199]: ERROR nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Traceback (most recent call last): [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self.driver.spawn(context, instance, image_meta, [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] vm_ref = self.build_virtual_machine(instance, [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.751005] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] for vif in network_info: [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return self._sync_wrapper(fn, *args, **kwargs) [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self.wait() [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self[:] = self._gt.wait() [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return self._exit_event.wait() [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] result = hub.switch() [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.751382] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return self.greenlet.switch() [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] result = function(*args, **kwargs) [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] return func(*args, **kwargs) [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] raise e [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] nwinfo = self.network_api.allocate_for_instance( [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] created_port_ids = self._update_ports_for_instance( [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] with excutils.save_and_reraise_exception(): [ 544.751715] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] self.force_reraise() [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] raise self.value [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] updated_port = self._update_port( [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] _ensure_no_port_binding_failure(port) [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] raise exception.PortBindingFailed(port_id=port['id']) [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] nova.exception.PortBindingFailed: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. [ 544.752077] env[62199]: ERROR nova.compute.manager [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] [ 544.752435] env[62199]: DEBUG nova.compute.utils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. {{(pid=62199) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.753006] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.654s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.754487] env[62199]: INFO nova.compute.claims [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.756137] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg a88326d154ea42ef83022b9c8550260f in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 544.757919] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Build of instance e34b059b-8edf-4a81-9485-358cefd2bde6 was re-scheduled: Binding failed for port da4c2f45-8714-4d2c-8997-38c3c011820b, please check neutron logs for more information. {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 544.758759] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Unplugging VIFs for instance {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 544.758992] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Acquiring lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.759135] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Acquired lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.759291] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.759902] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg f0d1765f86eb4e16884ab3a47ec4e3b6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 544.766669] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f0d1765f86eb4e16884ab3a47ec4e3b6 [ 544.771786] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Received event network-changed-446591f2-cdd1-4d8e-9486-e52602fe996e {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 544.772051] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Refreshing instance network info cache due to event network-changed-446591f2-cdd1-4d8e-9486-e52602fe996e. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 544.772722] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Acquiring lock "refresh_cache-52bd7ff6-1149-4a56-95c2-806572b526f4" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.772722] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Acquired lock "refresh_cache-52bd7ff6-1149-4a56-95c2-806572b526f4" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.772722] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Refreshing network info cache for port 446591f2-cdd1-4d8e-9486-e52602fe996e {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 544.773178] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Expecting reply to msg 58ae9d5ea3904dc099e6a570cd1b262f in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 544.783527] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 58ae9d5ea3904dc099e6a570cd1b262f [ 544.815033] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a88326d154ea42ef83022b9c8550260f [ 544.816878] env[62199]: DEBUG nova.compute.manager [None req-177dc621-e73e-4446-a33e-6f58a72c1b85 tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] [instance: a3aa8b80-5a13-4760-9288-25a05ea89442] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 544.818935] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-177dc621-e73e-4446-a33e-6f58a72c1b85 tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Expecting reply to msg 700a02c5085344d4892fbd3a8103df33 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 544.856165] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.856559] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Processing image 9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 544.856665] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.856805] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.856978] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 544.857226] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-effe332e-e1d7-40f3-95e9-11cac665a604 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.876852] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 544.877211] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62199) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 544.878144] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1056ad1-b061-4b07-99d0-943654d167b1 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.887448] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-528e2c03-6dc9-4d23-abef-46c392fcfb88 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.893370] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 544.893370] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]52d1419a-e3d6-0258-b51a-c0a3b61aa6f9" [ 544.893370] env[62199]: _type = "Task" [ 544.893370] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.899008] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 700a02c5085344d4892fbd3a8103df33 [ 544.918137] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Preparing fetch location {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 544.918377] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating directory with path [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 544.918609] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19635136-625e-4f5b-8788-6f76d33dd118 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.940964] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Created directory with path [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 544.941214] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Fetch image to [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 544.941325] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Downloading image file data 9fa830b9-8c21-4a7d-b7cb-64c85631deef to [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk on the data store datastore2 {{(pid=62199) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 544.942173] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09e8212-1adc-4a49-ad18-285c7aba9e43 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.949498] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86a1874e-5368-4c2f-9cf6-235c0a3abd1b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.958483] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd14412-2442-4de7-a921-0be3fde197b8 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.994582] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71d0a1c7-9f43-4c90-873b-aa2728108415 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.003641] env[62199]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-0e492779-4024-4c7b-94e9-37239e6913f9 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.031937] env[62199]: DEBUG nova.virt.vmwareapi.images [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Downloading image file data 9fa830b9-8c21-4a7d-b7cb-64c85631deef to the data store datastore2 {{(pid=62199) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 545.092278] env[62199]: DEBUG oslo_vmware.rw_handles [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62199) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 545.263973] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 9d1d24dab98845379a9ea3726faba0d2 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 545.272687] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9d1d24dab98845379a9ea3726faba0d2 [ 545.318791] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.327366] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.346487] env[62199]: DEBUG oslo_concurrency.lockutils [None req-177dc621-e73e-4446-a33e-6f58a72c1b85 tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.439586] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.440263] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Expecting reply to msg 00d5777427714268b283830a62f0f67d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 545.450607] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 00d5777427714268b283830a62f0f67d [ 545.527940] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.528483] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 3a38d99730e74c31b89f7a985c7d6a63 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 545.538970] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3a38d99730e74c31b89f7a985c7d6a63 [ 545.785955] env[62199]: DEBUG oslo_vmware.rw_handles [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Completed reading data from the image iterator. {{(pid=62199) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 545.786199] env[62199]: DEBUG oslo_vmware.rw_handles [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62199) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 545.948684] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Releasing lock "refresh_cache-52bd7ff6-1149-4a56-95c2-806572b526f4" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.948931] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Received event network-changed-3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 545.949087] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Refreshing instance network info cache due to event network-changed-3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 545.949372] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Acquiring lock "refresh_cache-a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 545.949530] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Acquired lock "refresh_cache-a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 545.949683] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Refreshing network info cache for port 3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 545.950171] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Expecting reply to msg 78063b27b3a24c24a042d5ba3187e729 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 545.951686] env[62199]: DEBUG nova.virt.vmwareapi.images [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Downloaded image file data 9fa830b9-8c21-4a7d-b7cb-64c85631deef to vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk on the data store datastore2 {{(pid=62199) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 545.953698] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Caching image {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 545.953867] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copying Virtual Disk [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk to [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 545.954352] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cc2e609d-0b19-474c-884f-81b047425d38 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.960242] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 78063b27b3a24c24a042d5ba3187e729 [ 545.962689] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 545.962689] env[62199]: value = "task-398568" [ 545.962689] env[62199]: _type = "Task" [ 545.962689] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 545.978088] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398568, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.030703] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Releasing lock "refresh_cache-e34b059b-8edf-4a81-9485-358cefd2bde6" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.030937] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 546.031120] env[62199]: DEBUG nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 546.031344] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.063170] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcce6d57-07be-4e1f-8044-aa9adc802489 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.067114] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.068083] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 3e678c70e6c14d1eaca65d9fcf268078 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 546.072976] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3ef5534-5c8e-4bfb-919e-2a5489f4668f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.078349] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3e678c70e6c14d1eaca65d9fcf268078 [ 546.107256] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742f089a-aabe-4521-aa37-fb57b47e70fe {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.116301] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561f582d-9b4b-4efd-bb0a-6d99fc4b302d {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.133038] env[62199]: DEBUG nova.compute.provider_tree [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.133645] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 43ac42d8a4cd45fa90f5226e351e4dd6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 546.142218] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 43ac42d8a4cd45fa90f5226e351e4dd6 [ 546.472399] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398568, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.573206] env[62199]: DEBUG nova.network.neutron [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.573968] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 63433d23947942f69247926df2fff5dc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 546.583733] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 63433d23947942f69247926df2fff5dc [ 546.636823] env[62199]: DEBUG nova.scheduler.client.report [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.639591] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 3afebf7f38e74eaa8fc72841b8c7f673 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 546.641639] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.653704] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3afebf7f38e74eaa8fc72841b8c7f673 [ 546.973494] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398568, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.693189} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 546.973814] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copied Virtual Disk [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk to [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 546.973944] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleting the datastore file [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef/tmp-sparse.vmdk {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 546.974265] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-77be7e14-25ad-47a0-8bf5-acf310fc7799 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.981162] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 546.981162] env[62199]: value = "task-398569" [ 546.981162] env[62199]: _type = "Task" [ 546.981162] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.989798] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398569, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.015717] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.016327] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Expecting reply to msg 29185ea4097d4a01bda62b8cadbdf6f3 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.026929] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 29185ea4097d4a01bda62b8cadbdf6f3 [ 547.076479] env[62199]: INFO nova.compute.manager [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] [instance: e34b059b-8edf-4a81-9485-358cefd2bde6] Took 1.04 seconds to deallocate network for instance. [ 547.078637] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg e5e0c302822c4de68d14523565fce344 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.118224] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e5e0c302822c4de68d14523565fce344 [ 547.145545] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 547.145995] env[62199]: DEBUG nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 547.147725] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 6a7317bfa7514469982abba3ee0e64fb in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.148661] env[62199]: DEBUG oslo_concurrency.lockutils [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 8.584s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.149445] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 6f5f573c59644fd193dfc0c99890c3ae in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.160146] env[62199]: ERROR nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 28f332db-d316-4659-a57c-3180ea938864, please check neutron logs for more information. [ 547.160146] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 547.160146] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.160146] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 547.160146] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.160146] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 547.160146] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.160146] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 547.160146] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.160146] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 547.160146] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.160146] env[62199]: ERROR nova.compute.manager raise self.value [ 547.160146] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.160146] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 547.160146] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.160146] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 547.160647] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.160647] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 547.160647] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 28f332db-d316-4659-a57c-3180ea938864, please check neutron logs for more information. [ 547.160647] env[62199]: ERROR nova.compute.manager [ 547.160647] env[62199]: Traceback (most recent call last): [ 547.160647] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 547.160647] env[62199]: listener.cb(fileno) [ 547.160647] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.160647] env[62199]: result = function(*args, **kwargs) [ 547.160647] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.160647] env[62199]: return func(*args, **kwargs) [ 547.160647] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.160647] env[62199]: raise e [ 547.160647] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.160647] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 547.160647] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.160647] env[62199]: created_port_ids = self._update_ports_for_instance( [ 547.160647] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.160647] env[62199]: with excutils.save_and_reraise_exception(): [ 547.160647] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.160647] env[62199]: self.force_reraise() [ 547.160647] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.160647] env[62199]: raise self.value [ 547.160647] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.160647] env[62199]: updated_port = self._update_port( [ 547.160647] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.160647] env[62199]: _ensure_no_port_binding_failure(port) [ 547.160647] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.160647] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 547.161508] env[62199]: nova.exception.PortBindingFailed: Binding failed for port 28f332db-d316-4659-a57c-3180ea938864, please check neutron logs for more information. [ 547.161508] env[62199]: Removing descriptor: 20 [ 547.161508] env[62199]: ERROR nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 28f332db-d316-4659-a57c-3180ea938864, please check neutron logs for more information. [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Traceback (most recent call last): [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] yield resources [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] self.driver.spawn(context, instance, image_meta, [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 547.161508] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] vm_ref = self.build_virtual_machine(instance, [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] vif_infos = vmwarevif.get_vif_info(self._session, [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] for vif in network_info: [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] return self._sync_wrapper(fn, *args, **kwargs) [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] self.wait() [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] self[:] = self._gt.wait() [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] return self._exit_event.wait() [ 547.161853] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] result = hub.switch() [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] return self.greenlet.switch() [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] result = function(*args, **kwargs) [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] return func(*args, **kwargs) [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] raise e [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] nwinfo = self.network_api.allocate_for_instance( [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 547.162255] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] created_port_ids = self._update_ports_for_instance( [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] with excutils.save_and_reraise_exception(): [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] self.force_reraise() [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] raise self.value [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] updated_port = self._update_port( [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] _ensure_no_port_binding_failure(port) [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 547.162616] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] raise exception.PortBindingFailed(port_id=port['id']) [ 547.162952] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] nova.exception.PortBindingFailed: Binding failed for port 28f332db-d316-4659-a57c-3180ea938864, please check neutron logs for more information. [ 547.162952] env[62199]: ERROR nova.compute.manager [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] [ 547.162952] env[62199]: INFO nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Terminating instance [ 547.163528] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Acquiring lock "refresh_cache-11e98624-f84b-420e-b5e1-42dbdd4779e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.163731] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Acquired lock "refresh_cache-11e98624-f84b-420e-b5e1-42dbdd4779e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.163909] env[62199]: DEBUG nova.network.neutron [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 547.164347] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg af3769a0b01449809cf2196c6c863d75 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.187593] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg af3769a0b01449809cf2196c6c863d75 [ 547.193630] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6f5f573c59644fd193dfc0c99890c3ae [ 547.202897] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6a7317bfa7514469982abba3ee0e64fb [ 547.491114] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398569, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.034116} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.491364] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleted the datastore file {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 547.491569] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Moving file from [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7/9fa830b9-8c21-4a7d-b7cb-64c85631deef to [datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef. {{(pid=62199) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 547.491816] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a2dcda82-46a5-4e32-be3a-c87803d8889c {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.500978] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 547.500978] env[62199]: value = "task-398570" [ 547.500978] env[62199]: _type = "Task" [ 547.500978] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.512747] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398570, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.526416] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Releasing lock "refresh_cache-a9cc78e1-979e-4b91-9ad1-31c5c5b342f7" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.526729] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 52bd7ff6-1149-4a56-95c2-806572b526f4] Received event network-vif-deleted-446591f2-cdd1-4d8e-9486-e52602fe996e {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 547.526938] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: a9cc78e1-979e-4b91-9ad1-31c5c5b342f7] Received event network-vif-deleted-3dfd82cf-a47f-45e8-8468-8a2eb4a06f0f {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 547.527124] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Received event network-changed-d17af2cc-d944-48ca-8570-36cc3677dc44 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 547.527285] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Refreshing instance network info cache due to event network-changed-d17af2cc-d944-48ca-8570-36cc3677dc44. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 547.527492] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Acquiring lock "refresh_cache-2681b401-47b3-4b05-b923-65e3078074d5" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.527642] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Acquired lock "refresh_cache-2681b401-47b3-4b05-b923-65e3078074d5" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.527878] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Refreshing network info cache for port d17af2cc-d944-48ca-8570-36cc3677dc44 {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 547.528445] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Expecting reply to msg 77443967c6fc44b69aa3b8e8e21fc48c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.540414] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 77443967c6fc44b69aa3b8e8e21fc48c [ 547.582909] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg 0a63edabe3234114b4c339c8272eb95f in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.618375] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0a63edabe3234114b4c339c8272eb95f [ 547.655218] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 4a4caef487b74b398ab299c0aee7a7ae in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.660129] env[62199]: DEBUG nova.compute.utils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.660129] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 5629e9dff9894ee8a3c37aa60d9bcf8e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.663858] env[62199]: DEBUG nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.664122] env[62199]: DEBUG nova.network.neutron [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 547.673611] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5629e9dff9894ee8a3c37aa60d9bcf8e [ 547.678813] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4a4caef487b74b398ab299c0aee7a7ae [ 547.713090] env[62199]: DEBUG nova.network.neutron [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.810688] env[62199]: DEBUG nova.policy [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '22941a93136b4b24ad702de3d44f6c0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2868e22d3af64ad4a663743b6c65c953', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 547.881292] env[62199]: DEBUG nova.network.neutron [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.881831] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Expecting reply to msg 7af9a1df876c4e99a3671745e37b481b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 547.891141] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7af9a1df876c4e99a3671745e37b481b [ 548.012308] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398570, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.031094} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.012308] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] File moved {{(pid=62199) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 548.012662] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Cleaning up location [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7 {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 548.012662] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleting the datastore file [datastore2] vmware_temp/a9066189-3ee2-43e4-ad99-05bd8558c0e7 {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 548.012862] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0588b98d-54c6-4b68-99a7-f98407beb2b4 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.019117] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 548.019117] env[62199]: value = "task-398571" [ 548.019117] env[62199]: _type = "Task" [ 548.019117] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.028522] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398571, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.097396] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.113945] env[62199]: INFO nova.scheduler.client.report [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Deleted allocations for instance e34b059b-8edf-4a81-9485-358cefd2bde6 [ 548.120459] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Expecting reply to msg b8a7ac391e134d74a5ca057969051097 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.136125] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b8a7ac391e134d74a5ca057969051097 [ 548.164813] env[62199]: DEBUG nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 548.166546] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 2875e89517c944dc9cec0e8605c77c0d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.187065] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 47b420a2-5eeb-4f14-9137-fb591a9dcbe6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.187212] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.187331] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.187447] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 52bd7ff6-1149-4a56-95c2-806572b526f4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.187558] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance a9cc78e1-979e-4b91-9ad1-31c5c5b342f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.187670] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 2681b401-47b3-4b05-b923-65e3078074d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.187813] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 11e98624-f84b-420e-b5e1-42dbdd4779e3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.187964] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 0840bc67-7390-4977-8004-f7a09358990d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 548.188557] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg f635adb3031a42f9a18b93b90f344693 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.218869] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2875e89517c944dc9cec0e8605c77c0d [ 548.219471] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f635adb3031a42f9a18b93b90f344693 [ 548.301194] env[62199]: DEBUG nova.compute.manager [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Received event network-changed-28f332db-d316-4659-a57c-3180ea938864 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 548.301369] env[62199]: DEBUG nova.compute.manager [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Refreshing instance network info cache due to event network-changed-28f332db-d316-4659-a57c-3180ea938864. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 548.301578] env[62199]: DEBUG oslo_concurrency.lockutils [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] Acquiring lock "refresh_cache-11e98624-f84b-420e-b5e1-42dbdd4779e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.325509] env[62199]: DEBUG nova.network.neutron [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.326013] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Expecting reply to msg 6aca3a48a22647dc8a3f7aaeb7519137 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.338280] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6aca3a48a22647dc8a3f7aaeb7519137 [ 548.384827] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Releasing lock "refresh_cache-11e98624-f84b-420e-b5e1-42dbdd4779e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.385563] env[62199]: DEBUG nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 548.385563] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 548.385689] env[62199]: DEBUG oslo_concurrency.lockutils [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] Acquired lock "refresh_cache-11e98624-f84b-420e-b5e1-42dbdd4779e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.385852] env[62199]: DEBUG nova.network.neutron [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Refreshing network info cache for port 28f332db-d316-4659-a57c-3180ea938864 {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 548.386273] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] Expecting reply to msg 42bb46c6e6cf41869fc716e759575245 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.387142] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6f995e57-673d-4a7e-8d7d-5115db8a26f2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.394828] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 42bb46c6e6cf41869fc716e759575245 [ 548.397877] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f966e430-3684-4b05-914e-339cfee20aa7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.421161] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 11e98624-f84b-420e-b5e1-42dbdd4779e3 could not be found. [ 548.421394] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 548.421573] env[62199]: INFO nova.compute.manager [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 548.421818] env[62199]: DEBUG oslo.service.loopingcall [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.422027] env[62199]: DEBUG nova.compute.manager [-] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 548.422118] env[62199]: DEBUG nova.network.neutron [-] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.483612] env[62199]: DEBUG nova.network.neutron [-] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.484151] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg f0d0d850e8d941eea55ee9db9aecaf78 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.490522] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f0d0d850e8d941eea55ee9db9aecaf78 [ 548.536189] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398571, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024277} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 548.536189] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Deleted the datastore file {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 548.536189] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e1117669-7475-4096-bf68-b03bcd8a248e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.537004] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 548.537004] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]52b2cedc-a014-9fc2-81c5-ab17fa30d54f" [ 548.537004] env[62199]: _type = "Task" [ 548.537004] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.548646] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52b2cedc-a014-9fc2-81c5-ab17fa30d54f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.623174] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9344d190-d79d-4593-a232-bced33a21ba7 tempest-ServerPasswordTestJSON-1217104794 tempest-ServerPasswordTestJSON-1217104794-project-member] Lock "e34b059b-8edf-4a81-9485-358cefd2bde6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 32.213s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.623604] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-4eb1ac2e-b373-4132-9b6d-09898ff087ed tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Expecting reply to msg dd4912bb06124a4dad34ff0fb37a51a5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.645573] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg dd4912bb06124a4dad34ff0fb37a51a5 [ 548.673185] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 8c6e211578af43949fc47c7791fde0a7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.693946] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 7a0241f3-41cc-494a-9a16-bf95b8abfab9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 548.694574] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 21f066c64e904e17846fba0288bf5b29 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.701814] env[62199]: DEBUG nova.network.neutron [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Successfully created port: 94e745ca-9bc7-4813-b516-d44f55d2813a {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.710417] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 21f066c64e904e17846fba0288bf5b29 [ 548.728193] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8c6e211578af43949fc47c7791fde0a7 [ 548.828896] env[62199]: DEBUG oslo_concurrency.lockutils [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] Releasing lock "refresh_cache-2681b401-47b3-4b05-b923-65e3078074d5" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 548.829185] env[62199]: DEBUG nova.compute.manager [req-740d4ec3-a80d-4c53-9e69-a13a6c6b3385 req-c7409df7-7c2f-4cb0-992b-7bd46ee73c2d service nova] [instance: 2681b401-47b3-4b05-b923-65e3078074d5] Received event network-vif-deleted-d17af2cc-d944-48ca-8570-36cc3677dc44 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 548.919554] env[62199]: DEBUG nova.network.neutron [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.986455] env[62199]: DEBUG nova.network.neutron [-] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.987039] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 85958db6652d44a58617311c8f1de62e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 548.997260] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 85958db6652d44a58617311c8f1de62e [ 549.048787] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52b2cedc-a014-9fc2-81c5-ab17fa30d54f, 'name': SearchDatastore_Task, 'duration_secs': 0.027616} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.049095] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.049341] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore2] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 549.049541] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-406a6099-a2ef-4099-9231-610bdc2ee018 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.056203] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 549.056203] env[62199]: value = "task-398572" [ 549.056203] env[62199]: _type = "Task" [ 549.056203] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.065186] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398572, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.126186] env[62199]: DEBUG nova.compute.manager [None req-4eb1ac2e-b373-4132-9b6d-09898ff087ed tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] [instance: 4ddf6b0c-f752-4080-bdef-70d954d2c4f6] Starting instance... {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 549.128040] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-4eb1ac2e-b373-4132-9b6d-09898ff087ed tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Expecting reply to msg e82881450270454b9260d9575890f8b7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 549.169299] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e82881450270454b9260d9575890f8b7 [ 549.175410] env[62199]: DEBUG nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 549.198783] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 549.199361] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 7084bdc55abb4a80be914aeeaaf2028e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 549.204558] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.204558] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.204558] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.205226] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.205226] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.205226] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.205226] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.205226] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.205490] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.205490] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.205490] env[62199]: DEBUG nova.virt.hardware [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.206621] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3f6cac-dee3-4d45-88f0-f941e7f0f9ab {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.215385] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb37bd3-ee93-4e2f-b3ec-f89ab8503090 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.265341] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7084bdc55abb4a80be914aeeaaf2028e [ 549.274173] env[62199]: DEBUG nova.network.neutron [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.274700] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] Expecting reply to msg 69db95d6c49144b1afa098a39f805f91 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 549.283970] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 69db95d6c49144b1afa098a39f805f91 [ 549.489215] env[62199]: INFO nova.compute.manager [-] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Took 1.07 seconds to deallocate network for instance. [ 549.491949] env[62199]: DEBUG nova.compute.claims [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 549.492188] env[62199]: DEBUG oslo_concurrency.lockutils [None req-39bd0b20-cda9-4399-9dfd-113df1eef442 tempest-MigrationsAdminTest-827511960 tempest-MigrationsAdminTest-827511960-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.566173] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398572, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435242} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 549.566460] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore2] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 549.566680] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Extending root virtual disk to 1048576 {{(pid=62199) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 549.566960] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11134cca-b3b1-4a11-a808-6cdd1854080e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.574518] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 549.574518] env[62199]: value = "task-398573" [ 549.574518] env[62199]: _type = "Task" [ 549.574518] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.582629] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398573, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.646060] env[62199]: DEBUG oslo_concurrency.lockutils [None req-4eb1ac2e-b373-4132-9b6d-09898ff087ed tempest-ListImageFiltersTestJSON-1753121278 tempest-ListImageFiltersTestJSON-1753121278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.702153] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance a3aa8b80-5a13-4760-9288-25a05ea89442 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 549.702813] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 9983b591b7364653ab83489a202203cc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 549.716242] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9983b591b7364653ab83489a202203cc [ 549.777209] env[62199]: DEBUG oslo_concurrency.lockutils [req-4f1f325e-2fcc-4c75-b811-7c699a7a262d req-1db03f12-63b1-435a-b9c0-758355942b4f service nova] Releasing lock "refresh_cache-11e98624-f84b-420e-b5e1-42dbdd4779e3" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.080958] env[62199]: DEBUG oslo_concurrency.lockutils [None req-01e770ac-2a4f-4315-9e9d-b742d0619914 tempest-ServerDiagnosticsNegativeTest-1300502770 tempest-ServerDiagnosticsNegativeTest-1300502770-project-member] Acquiring lock "56c98167-585f-4e33-8119-59a0c2798ee8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.081231] env[62199]: DEBUG oslo_concurrency.lockutils [None req-01e770ac-2a4f-4315-9e9d-b742d0619914 tempest-ServerDiagnosticsNegativeTest-1300502770 tempest-ServerDiagnosticsNegativeTest-1300502770-project-member] Lock "56c98167-585f-4e33-8119-59a0c2798ee8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.086972] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398573, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069107} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.087220] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Extended root virtual disk {{(pid=62199) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 550.088066] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b51b31-5e19-47ea-93b4-cabc239ce11b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.109396] env[62199]: DEBUG nova.virt.vmwareapi.volumeops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Reconfiguring VM instance instance-00000003 to attach disk [datastore2] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk or device None with type sparse {{(pid=62199) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 550.109708] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8245aff7-b9ff-418a-891c-d90596f87c89 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.129977] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 550.129977] env[62199]: value = "task-398574" [ 550.129977] env[62199]: _type = "Task" [ 550.129977] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.207161] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 4ddf6b0c-f752-4080-bdef-70d954d2c4f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 550.207161] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 71fd6033268f4197817f7bd7c6e474d7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 550.223384] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 71fd6033268f4197817f7bd7c6e474d7 [ 550.640883] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398574, 'name': ReconfigVM_Task, 'duration_secs': 0.406044} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.641218] env[62199]: DEBUG nova.virt.vmwareapi.volumeops [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Reconfigured VM instance instance-00000003 to attach disk [datastore2] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk or device None with type sparse {{(pid=62199) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 550.641822] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-22516de6-869b-4ee5-a1d5-100e490eca8b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.648267] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 550.648267] env[62199]: value = "task-398575" [ 550.648267] env[62199]: _type = "Task" [ 550.648267] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.660052] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398575, 'name': Rename_Task} progress is 5%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.715082] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 4b41ffc1-acc3-4568-b34e-548a0880a30d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 550.716062] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 67bce4bc5acb4a4c8fedb7ccb1ae77de in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 550.734037] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 67bce4bc5acb4a4c8fedb7ccb1ae77de [ 551.158007] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398575, 'name': Rename_Task, 'duration_secs': 0.129711} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.158249] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powering on the VM {{(pid=62199) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 551.158490] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-123b225c-97ac-4699-b23a-0aa3473c034b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.165651] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Waiting for the task: (returnval){ [ 551.165651] env[62199]: value = "task-398576" [ 551.165651] env[62199]: _type = "Task" [ 551.165651] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.175376] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398576, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.218869] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance b56237fc-838d-4ce2-b435-60b4f611f113 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.219484] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 290604de7b7a47d38e48a45a3447b161 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 551.230622] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 290604de7b7a47d38e48a45a3447b161 [ 551.676877] env[62199]: DEBUG oslo_vmware.api [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Task: {'id': task-398576, 'name': PowerOnVM_Task, 'duration_secs': 0.436753} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.676877] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powered on the VM {{(pid=62199) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 551.676877] env[62199]: DEBUG nova.compute.manager [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Checking state {{(pid=62199) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 551.677614] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e3bdc1a-8943-4fab-a87a-9aa776055ce5 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.685631] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Expecting reply to msg b33df688e21549128f6a73e64cf4ebe5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 551.722530] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance ccaed194-e8b4-4620-be97-588a232af696 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 551.723182] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 29049a09091f43d3ae2f92a1863a92ca in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 551.737716] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 29049a09091f43d3ae2f92a1863a92ca [ 551.770886] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b33df688e21549128f6a73e64cf4ebe5 [ 551.798980] env[62199]: ERROR nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 94e745ca-9bc7-4813-b516-d44f55d2813a, please check neutron logs for more information. [ 551.798980] env[62199]: ERROR nova.compute.manager Traceback (most recent call last): [ 551.798980] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.798980] env[62199]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 551.798980] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.798980] env[62199]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 551.798980] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.798980] env[62199]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 551.798980] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.798980] env[62199]: ERROR nova.compute.manager self.force_reraise() [ 551.798980] env[62199]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.798980] env[62199]: ERROR nova.compute.manager raise self.value [ 551.798980] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.798980] env[62199]: ERROR nova.compute.manager updated_port = self._update_port( [ 551.798980] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.798980] env[62199]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 551.799473] env[62199]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.799473] env[62199]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 551.799473] env[62199]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 94e745ca-9bc7-4813-b516-d44f55d2813a, please check neutron logs for more information. [ 551.799473] env[62199]: ERROR nova.compute.manager [ 551.799473] env[62199]: Traceback (most recent call last): [ 551.799473] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 551.799473] env[62199]: listener.cb(fileno) [ 551.799473] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.799473] env[62199]: result = function(*args, **kwargs) [ 551.799473] env[62199]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.799473] env[62199]: return func(*args, **kwargs) [ 551.799473] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.799473] env[62199]: raise e [ 551.799473] env[62199]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.799473] env[62199]: nwinfo = self.network_api.allocate_for_instance( [ 551.799473] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.799473] env[62199]: created_port_ids = self._update_ports_for_instance( [ 551.799473] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.799473] env[62199]: with excutils.save_and_reraise_exception(): [ 551.799473] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.799473] env[62199]: self.force_reraise() [ 551.799473] env[62199]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.799473] env[62199]: raise self.value [ 551.799473] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.799473] env[62199]: updated_port = self._update_port( [ 551.799473] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.799473] env[62199]: _ensure_no_port_binding_failure(port) [ 551.799473] env[62199]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.799473] env[62199]: raise exception.PortBindingFailed(port_id=port['id']) [ 551.800395] env[62199]: nova.exception.PortBindingFailed: Binding failed for port 94e745ca-9bc7-4813-b516-d44f55d2813a, please check neutron logs for more information. [ 551.800395] env[62199]: Removing descriptor: 20 [ 551.800395] env[62199]: ERROR nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 94e745ca-9bc7-4813-b516-d44f55d2813a, please check neutron logs for more information. [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] Traceback (most recent call last): [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] yield resources [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] self.driver.spawn(context, instance, image_meta, [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 551.800395] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] vm_ref = self.build_virtual_machine(instance, [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] vif_infos = vmwarevif.get_vif_info(self._session, [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] for vif in network_info: [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] return self._sync_wrapper(fn, *args, **kwargs) [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] self.wait() [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] self[:] = self._gt.wait() [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] return self._exit_event.wait() [ 551.800955] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] result = hub.switch() [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] return self.greenlet.switch() [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] result = function(*args, **kwargs) [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] return func(*args, **kwargs) [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] raise e [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] nwinfo = self.network_api.allocate_for_instance( [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 551.801384] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] created_port_ids = self._update_ports_for_instance( [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] with excutils.save_and_reraise_exception(): [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] self.force_reraise() [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] raise self.value [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] updated_port = self._update_port( [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] _ensure_no_port_binding_failure(port) [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 551.801753] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] raise exception.PortBindingFailed(port_id=port['id']) [ 551.802137] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] nova.exception.PortBindingFailed: Binding failed for port 94e745ca-9bc7-4813-b516-d44f55d2813a, please check neutron logs for more information. [ 551.802137] env[62199]: ERROR nova.compute.manager [instance: 0840bc67-7390-4977-8004-f7a09358990d] [ 551.802137] env[62199]: INFO nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Terminating instance [ 551.803865] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Acquiring lock "refresh_cache-0840bc67-7390-4977-8004-f7a09358990d" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 551.803865] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Acquired lock "refresh_cache-0840bc67-7390-4977-8004-f7a09358990d" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 551.803865] env[62199]: DEBUG nova.network.neutron [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 551.804609] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 7197c0d120bb4f6ca5a5be4ca0cf0f21 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 551.810467] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7197c0d120bb4f6ca5a5be4ca0cf0f21 [ 551.863326] env[62199]: DEBUG nova.compute.manager [req-5d605057-d873-41c3-80c8-f048d0707e0e req-b946edc4-1357-477c-8511-8befd852bc1b service nova] [instance: 11e98624-f84b-420e-b5e1-42dbdd4779e3] Received event network-vif-deleted-28f332db-d316-4659-a57c-3180ea938864 {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 552.195854] env[62199]: DEBUG oslo_concurrency.lockutils [None req-052c21f7-54b1-4477-8629-510e47b4d176 tempest-ServersAdmin275Test-185658888 tempest-ServersAdmin275Test-185658888-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.225535] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance e047a038-bcf1-4f90-929a-465bbee3ef67 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.226077] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 2c9b4714f7a54ff4beaa6f5083f7a293 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 552.238392] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2c9b4714f7a54ff4beaa6f5083f7a293 [ 552.342775] env[62199]: DEBUG nova.network.neutron [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 552.557763] env[62199]: DEBUG nova.network.neutron [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.558321] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Expecting reply to msg 3ca20a7c8d164b84990aa7f691e939fa in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 552.567699] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3ca20a7c8d164b84990aa7f691e939fa [ 552.733291] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Instance 77de9475-77f5-4833-aaa4-8a1d6de6db9b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62199) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 552.733615] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62199) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 552.733708] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62199) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 552.751668] env[62199]: DEBUG nova.compute.manager [None req-e086fc91-bdf7-4c1e-a38f-24870499d3e0 tempest-ServerDiagnosticsV248Test-1307978044 tempest-ServerDiagnosticsV248Test-1307978044-project-admin] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Checking state {{(pid=62199) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 552.753050] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c782f94b-327d-45e4-b530-603aab69f559 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.759999] env[62199]: INFO nova.compute.manager [None req-e086fc91-bdf7-4c1e-a38f-24870499d3e0 tempest-ServerDiagnosticsV248Test-1307978044 tempest-ServerDiagnosticsV248Test-1307978044-project-admin] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Retrieving diagnostics [ 552.763496] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa73e206-fdc6-4617-86b6-29ebe15dffca {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.009573] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-283426d3-e0ce-4b04-ad2f-c30a81388273 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.019413] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eae351d7-9d5a-45d8-ba8c-cc70378dda5b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.055345] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f42e9a-08f4-40f1-9f24-ee7bf163c465 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.060243] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Releasing lock "refresh_cache-0840bc67-7390-4977-8004-f7a09358990d" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.060643] env[62199]: DEBUG nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 553.061010] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 553.063249] env[62199]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-20115028-6277-4dd5-8e52-fc36b9822587 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.066248] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628c129e-92af-4151-ad55-1a64e0111988 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.082821] env[62199]: DEBUG nova.compute.provider_tree [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 553.083250] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 19c5efc8e0cd4556a111fc97eb8e8244 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 553.088300] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa91578-7915-478f-817e-c73ef470d5f2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.098579] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 19c5efc8e0cd4556a111fc97eb8e8244 [ 553.113444] env[62199]: WARNING nova.virt.vmwareapi.vmops [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0840bc67-7390-4977-8004-f7a09358990d could not be found. [ 553.113672] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 553.113841] env[62199]: INFO nova.compute.manager [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 553.114202] env[62199]: DEBUG oslo.service.loopingcall [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 553.114706] env[62199]: DEBUG nova.compute.manager [-] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 553.114706] env[62199]: DEBUG nova.network.neutron [-] [instance: 0840bc67-7390-4977-8004-f7a09358990d] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 553.130611] env[62199]: DEBUG nova.network.neutron [-] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.131164] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 3ca4b2682efb4e0ea48a1cf5c62ca41e in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 553.137984] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3ca4b2682efb4e0ea48a1cf5c62ca41e [ 553.587357] env[62199]: DEBUG nova.scheduler.client.report [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.589999] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Expecting reply to msg 699e8c9e685b4960a5b900520de27abc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 553.604116] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 699e8c9e685b4960a5b900520de27abc [ 553.632866] env[62199]: DEBUG nova.network.neutron [-] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.633398] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 27d90d2e9753498fa31bd1c8ef8d913c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 553.642379] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 27d90d2e9753498fa31bd1c8ef8d913c [ 554.092354] env[62199]: DEBUG nova.compute.resource_tracker [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62199) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 554.092612] env[62199]: DEBUG oslo_concurrency.lockutils [None req-84bf0c1d-b130-495b-9c3a-0108fcae14c3 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.944s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.092895] env[62199]: DEBUG oslo_concurrency.lockutils [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.849s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 554.094525] env[62199]: INFO nova.compute.claims [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 554.096115] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg ad9719709fb549c590977b380d2b07dc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 554.135211] env[62199]: INFO nova.compute.manager [-] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Took 1.02 seconds to deallocate network for instance. [ 554.137826] env[62199]: DEBUG nova.compute.claims [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Aborting claim: {{(pid=62199) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 554.138006] env[62199]: DEBUG oslo_concurrency.lockutils [None req-2dc7489f-278d-46d6-8b3a-49e76d27cfc0 tempest-ImagesOneServerTestJSON-2075344246 tempest-ImagesOneServerTestJSON-2075344246-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 554.139429] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ad9719709fb549c590977b380d2b07dc [ 554.601938] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 8725d6470c5241ceb4ae9db532741c01 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 554.614513] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8725d6470c5241ceb4ae9db532741c01 [ 555.155225] env[62199]: DEBUG oslo_concurrency.lockutils [None req-3ff90742-7844-4531-b77b-a383fffdd2c8 tempest-ServersTestMultiNic-83593086 tempest-ServersTestMultiNic-83593086-project-member] Acquiring lock "e5a91d8d-dec1-4026-a498-53e892121962" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.155517] env[62199]: DEBUG oslo_concurrency.lockutils [None req-3ff90742-7844-4531-b77b-a383fffdd2c8 tempest-ServersTestMultiNic-83593086 tempest-ServersTestMultiNic-83593086-project-member] Lock "e5a91d8d-dec1-4026-a498-53e892121962" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.278768] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 1dadf05f56cc42a19c57366dfc2c7ff0 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 555.298520] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1dadf05f56cc42a19c57366dfc2c7ff0 [ 555.409338] env[62199]: DEBUG oslo_concurrency.lockutils [None req-92b089f4-4ff8-4d4b-a838-1469a5c3461b tempest-AttachInterfacesUnderV243Test-755047540 tempest-AttachInterfacesUnderV243Test-755047540-project-member] Acquiring lock "a36af4bb-2c96-4d9c-9964-21aef231d695" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.409557] env[62199]: DEBUG oslo_concurrency.lockutils [None req-92b089f4-4ff8-4d4b-a838-1469a5c3461b tempest-AttachInterfacesUnderV243Test-755047540 tempest-AttachInterfacesUnderV243Test-755047540-project-member] Lock "a36af4bb-2c96-4d9c-9964-21aef231d695" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.447055] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40e3332c-da88-4ce1-90df-0d5c3196a025 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.454900] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b662f46c-37b1-43ac-9127-cb14455553b9 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.494941] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48dba643-f7b1-47a3-b305-865759bbfa1b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.504408] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d266ba-4ba4-4b59-9cbd-078dda0b600a {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.520355] env[62199]: DEBUG nova.compute.provider_tree [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 555.520859] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 41a114a5e1274270bdbbb9b8c8f99137 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 555.532123] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 41a114a5e1274270bdbbb9b8c8f99137 [ 555.781083] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.781351] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.782932] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.782932] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.782932] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Lock "83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.784995] env[62199]: INFO nova.compute.manager [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Terminating instance [ 555.787976] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "refresh_cache-83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.788152] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquired lock "refresh_cache-83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.788314] env[62199]: DEBUG nova.network.neutron [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 555.788754] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 63417ebdeecc437597505306dbbd2863 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 555.798957] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 63417ebdeecc437597505306dbbd2863 [ 556.025002] env[62199]: DEBUG nova.scheduler.client.report [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 556.027420] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 9a7501d9425c4c08bd2a736d8e7f0c60 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.040218] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9a7501d9425c4c08bd2a736d8e7f0c60 [ 556.126118] env[62199]: DEBUG nova.compute.manager [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Received event network-changed-94e745ca-9bc7-4813-b516-d44f55d2813a {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 556.126338] env[62199]: DEBUG nova.compute.manager [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Refreshing instance network info cache due to event network-changed-94e745ca-9bc7-4813-b516-d44f55d2813a. {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11134}} [ 556.126562] env[62199]: DEBUG oslo_concurrency.lockutils [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] Acquiring lock "refresh_cache-0840bc67-7390-4977-8004-f7a09358990d" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.126792] env[62199]: DEBUG oslo_concurrency.lockutils [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] Acquired lock "refresh_cache-0840bc67-7390-4977-8004-f7a09358990d" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.126908] env[62199]: DEBUG nova.network.neutron [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Refreshing network info cache for port 94e745ca-9bc7-4813-b516-d44f55d2813a {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 556.127351] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] Expecting reply to msg a1f5106c3fbe4b1689d1ab357c706d67 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.136780] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a1f5106c3fbe4b1689d1ab357c706d67 [ 556.148544] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Expecting reply to msg 2b61aee5202b4a12bf511ae44fdd4f97 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.171831] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2b61aee5202b4a12bf511ae44fdd4f97 [ 556.325433] env[62199]: DEBUG nova.network.neutron [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.376292] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d3058c88-d9a6-489a-9c1b-adad75b6a817 tempest-TenantUsagesTestJSON-979742911 tempest-TenantUsagesTestJSON-979742911-project-member] Acquiring lock "946a30b4-9574-4b9b-8bc7-e951c7c75a9b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.376822] env[62199]: DEBUG oslo_concurrency.lockutils [None req-d3058c88-d9a6-489a-9c1b-adad75b6a817 tempest-TenantUsagesTestJSON-979742911 tempest-TenantUsagesTestJSON-979742911-project-member] Lock "946a30b4-9574-4b9b-8bc7-e951c7c75a9b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.388268] env[62199]: DEBUG nova.network.neutron [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.388755] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 3ddcd87d326242cc9db7dd5f91f01c37 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.397355] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3ddcd87d326242cc9db7dd5f91f01c37 [ 556.530144] env[62199]: DEBUG oslo_concurrency.lockutils [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.437s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 556.530506] env[62199]: DEBUG nova.compute.manager [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 556.532372] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 7cf24fd91d844a3b89f000ddd3536e54 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.540104] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.448s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 556.540104] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg b0687c8fb22842519120a785578b44f7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.576947] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7cf24fd91d844a3b89f000ddd3536e54 [ 556.577794] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b0687c8fb22842519120a785578b44f7 [ 556.650345] env[62199]: INFO nova.compute.manager [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Rebuilding instance [ 556.709275] env[62199]: DEBUG nova.compute.manager [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Checking state {{(pid=62199) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 556.710316] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9d09b0b-3567-4794-afcf-b828c155b531 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.714434] env[62199]: DEBUG nova.network.neutron [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 556.723698] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Expecting reply to msg 07f15894eef247c89fc41bbb11bcaef1 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.783821] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 07f15894eef247c89fc41bbb11bcaef1 [ 556.895438] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Releasing lock "refresh_cache-83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.895438] env[62199]: DEBUG nova.compute.manager [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Start destroying the instance on the hypervisor. {{(pid=62199) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 556.895438] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 556.895438] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4346295-5037-4666-8484-778334f13abd {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.900513] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Powering off the VM {{(pid=62199) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 556.901352] env[62199]: DEBUG nova.network.neutron [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.902066] env[62199]: INFO oslo_messaging._drivers.amqpdriver [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] Expecting reply to msg 092d33ad1e9f43899795042890957729 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 556.904366] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b920353b-5197-4d26-8d9c-b70a893e1026 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.910521] env[62199]: DEBUG oslo_vmware.api [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 556.910521] env[62199]: value = "task-398582" [ 556.910521] env[62199]: _type = "Task" [ 556.910521] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.917528] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 092d33ad1e9f43899795042890957729 [ 556.924936] env[62199]: DEBUG oslo_vmware.api [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398582, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.038055] env[62199]: DEBUG nova.compute.utils [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 557.038732] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg ec056af83ac74901979b8f23f8a59a82 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 557.040161] env[62199]: DEBUG nova.compute.manager [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 557.040161] env[62199]: DEBUG nova.network.neutron [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 557.058835] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ec056af83ac74901979b8f23f8a59a82 [ 557.204996] env[62199]: DEBUG nova.policy [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cf2a6b14bf624f52a30408e0f0223b1e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2ee3d1f037784721a6360ac0f3cb565b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 557.229108] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powering off the VM {{(pid=62199) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 557.229592] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95a4955c-da3a-4dc8-9d6d-efd0afb270f7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.235594] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for the task: (returnval){ [ 557.235594] env[62199]: value = "task-398584" [ 557.235594] env[62199]: _type = "Task" [ 557.235594] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.250810] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': task-398584, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.371008] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c1a7085-b39f-42fd-a541-ddea041d56d8 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.381293] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f9cf8e-0b0b-4f51-8672-665013c1d29f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.417087] env[62199]: DEBUG oslo_concurrency.lockutils [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] Releasing lock "refresh_cache-0840bc67-7390-4977-8004-f7a09358990d" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 557.417322] env[62199]: DEBUG nova.compute.manager [req-3e1df3fe-7745-4847-91b8-2c0407382183 req-1cb8d359-924f-4e6f-b1c6-1f24fe51f68a service nova] [instance: 0840bc67-7390-4977-8004-f7a09358990d] Received event network-vif-deleted-94e745ca-9bc7-4813-b516-d44f55d2813a {{(pid=62199) external_instance_event /opt/stack/nova/nova/compute/manager.py:11129}} [ 557.422693] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8b7ba9-5062-4166-bf31-7ceb06db3c58 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.432359] env[62199]: DEBUG oslo_vmware.api [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398582, 'name': PowerOffVM_Task, 'duration_secs': 0.163411} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.435089] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Powered off the VM {{(pid=62199) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 557.435297] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Unregistering the VM {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 557.435589] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3e7dbbb9-0849-40c3-b989-ae25dfd9d060 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.438181] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177c4ecf-74d6-4ab3-83d0-788842baadf7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.453562] env[62199]: DEBUG nova.compute.provider_tree [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 557.453978] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 0b4dfee2d78c467ab2f9228a582277e2 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 557.463615] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0b4dfee2d78c467ab2f9228a582277e2 [ 557.470142] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Unregistered the VM {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 557.470142] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Deleting contents of the VM from datastore datastore1 {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 557.470142] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Deleting the datastore file [datastore1] 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8 {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 557.470643] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e65239a-4ce1-49c6-84fc-33047de38b75 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.476877] env[62199]: DEBUG oslo_vmware.api [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for the task: (returnval){ [ 557.476877] env[62199]: value = "task-398586" [ 557.476877] env[62199]: _type = "Task" [ 557.476877] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.486336] env[62199]: DEBUG oslo_vmware.api [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398586, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.555867] env[62199]: DEBUG nova.compute.manager [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 557.555867] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 910359bbf6b5468abd374f950ef7577c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 557.596932] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 910359bbf6b5468abd374f950ef7577c [ 557.747486] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': task-398584, 'name': PowerOffVM_Task, 'duration_secs': 0.119881} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.748035] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Powered off the VM {{(pid=62199) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 557.748402] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Destroying instance {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 557.749297] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c66eabe-12a2-44c7-b149-05ddcf64ea49 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.759048] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Unregistering the VM {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 557.759467] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-860ef790-6303-45f6-a29e-c87c33724582 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.788044] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Unregistered the VM {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 557.788044] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Deleting contents of the VM from datastore datastore2 {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 557.788044] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Deleting the datastore file [datastore2] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6 {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 557.788044] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a4189478-2a85-44b9-ac7c-6e6aec9b26b3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.802251] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for the task: (returnval){ [ 557.802251] env[62199]: value = "task-398588" [ 557.802251] env[62199]: _type = "Task" [ 557.802251] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.813256] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': task-398588, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.957134] env[62199]: DEBUG nova.scheduler.client.report [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 557.959679] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 05b80be303314d92ad9f56808425098c in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 557.982458] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 05b80be303314d92ad9f56808425098c [ 557.989854] env[62199]: DEBUG oslo_vmware.api [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Task: {'id': task-398586, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.09939} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.990345] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Deleted the datastore file {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 557.990345] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Deleted contents of the VM from datastore datastore1 {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 557.990468] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 557.990604] env[62199]: INFO nova.compute.manager [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Took 1.10 seconds to destroy the instance on the hypervisor. [ 557.990861] env[62199]: DEBUG oslo.service.loopingcall [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 557.991030] env[62199]: DEBUG nova.compute.manager [-] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 557.991126] env[62199]: DEBUG nova.network.neutron [-] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 558.060978] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Expecting reply to msg 8e71ec58fa41426a8042faeea275f713 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.085772] env[62199]: DEBUG nova.network.neutron [-] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.086288] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg a84812d690f14cb5944fe33a8e46edbc in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.095371] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a84812d690f14cb5944fe33a8e46edbc [ 558.098208] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8e71ec58fa41426a8042faeea275f713 [ 558.317586] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': task-398588, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10037} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.317906] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Deleted the datastore file {{(pid=62199) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 558.318156] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Deleted contents of the VM from datastore datastore2 {{(pid=62199) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 558.318331] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Instance destroyed {{(pid=62199) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 558.320104] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Expecting reply to msg 8b469872d9ec45f587114a9e8a548b7d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.355231] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8b469872d9ec45f587114a9e8a548b7d [ 558.462293] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.929s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.462944] env[62199]: ERROR nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Traceback (most recent call last): [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self.driver.spawn(context, instance, image_meta, [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] vm_ref = self.build_virtual_machine(instance, [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] vif_infos = vmwarevif.get_vif_info(self._session, [ 558.462944] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] for vif in network_info: [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return self._sync_wrapper(fn, *args, **kwargs) [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self.wait() [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self[:] = self._gt.wait() [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return self._exit_event.wait() [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] result = hub.switch() [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 558.463287] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return self.greenlet.switch() [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] result = function(*args, **kwargs) [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] return func(*args, **kwargs) [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] raise e [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] nwinfo = self.network_api.allocate_for_instance( [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] created_port_ids = self._update_ports_for_instance( [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] with excutils.save_and_reraise_exception(): [ 558.463720] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] self.force_reraise() [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] raise self.value [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] updated_port = self._update_port( [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] _ensure_no_port_binding_failure(port) [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] raise exception.PortBindingFailed(port_id=port['id']) [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] nova.exception.PortBindingFailed: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. [ 558.464470] env[62199]: ERROR nova.compute.manager [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] [ 558.465213] env[62199]: DEBUG nova.compute.utils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. {{(pid=62199) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 558.465213] env[62199]: DEBUG oslo_concurrency.lockutils [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.010s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.466747] env[62199]: INFO nova.compute.claims [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] [instance: 7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.468581] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg 49192ee14ee24cd49d48c549cf06a83a in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.480433] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Build of instance 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f was re-scheduled: Binding failed for port e7d41690-0430-4423-a444-832d87c671e3, please check neutron logs for more information. {{(pid=62199) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 558.480433] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Unplugging VIFs for instance {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 558.480433] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Acquiring lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.480433] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Acquired lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 558.481124] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Building network info cache for instance {{(pid=62199) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 558.481124] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 7ef20024e7264de788fd77fc800b8f4b in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.481355] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7ef20024e7264de788fd77fc800b8f4b [ 558.512795] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 49192ee14ee24cd49d48c549cf06a83a [ 558.563582] env[62199]: DEBUG nova.compute.manager [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] Start spawning the instance on the hypervisor. {{(pid=62199) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 558.588878] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:07:01Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='871814930',id=22,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1419056952',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 558.589206] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 558.589362] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 558.589539] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 558.589678] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 558.589898] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 558.590116] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 558.590272] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 558.590435] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 558.590591] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 558.590774] env[62199]: DEBUG nova.virt.hardware [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 558.591604] env[62199]: DEBUG nova.network.neutron [-] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.591886] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 199e92b0ef7a46e4a06776a1b31f6489 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.593475] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da906cb2-b705-4cea-a96f-4a03c8915fd7 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.601532] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 199e92b0ef7a46e4a06776a1b31f6489 [ 558.602955] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8bf557c-6400-472f-aa1c-be6f0ba67eb2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.825773] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Expecting reply to msg 36e172ad08a94c62b6b76b11efdb42ea in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.854384] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 36e172ad08a94c62b6b76b11efdb42ea [ 558.974942] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg 067bc16d9c884fb3bdbee78da8f1bebd in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 558.989036] env[62199]: DEBUG nova.network.neutron [None req-c6709696-f8c0-4aa8-9fdb-d3b0542d7208 tempest-ServersWithSpecificFlavorTestJSON-30428226 tempest-ServersWithSpecificFlavorTestJSON-30428226-project-member] [instance: 7a0241f3-41cc-494a-9a16-bf95b8abfab9] Successfully created port: fe09ff9d-06fa-4c98-91b1-b314fe0e7597 {{(pid=62199) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 558.992036] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 067bc16d9c884fb3bdbee78da8f1bebd [ 559.084581] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.097084] env[62199]: INFO nova.compute.manager [-] [instance: 83fa93bc-04b9-4ace-aeaf-7dbcd3c456d8] Took 1.11 seconds to deallocate network for instance. [ 559.102428] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Expecting reply to msg 935f82e8a7554feead9c784ce9257f71 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 559.139175] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 935f82e8a7554feead9c784ce9257f71 [ 559.350438] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-08-19T23:05:12Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-19T23:04:54Z,direct_url=,disk_format='vmdk',id=9fa830b9-8c21-4a7d-b7cb-64c85631deef,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='b4e002c64ce746f8a9110b954b989005',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-19T23:04:55Z,virtual_size=,visibility=), allow threads: False {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 559.350697] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Flavor limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 559.350981] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Image limits 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 559.351081] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Flavor pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 559.351209] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Image pref 0:0:0 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 559.351356] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62199) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 559.351556] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 559.351741] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 559.352073] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Got 1 possible topologies {{(pid=62199) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 559.352073] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 559.353329] env[62199]: DEBUG nova.virt.hardware [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62199) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 559.353329] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c0c4656-3a49-4f3c-b539-03735f5c18a5 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.360963] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85831c92-56b2-4ad5-b624-eea0d8ba00ad {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.375059] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Instance VIF info [] {{(pid=62199) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 559.381908] env[62199]: DEBUG oslo.service.loopingcall [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62199) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.382205] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Creating VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 559.382435] env[62199]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-663cb5bb-55be-4d15-bcbb-a034a922160b {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.399936] env[62199]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 559.399936] env[62199]: value = "task-398590" [ 559.399936] env[62199]: _type = "Task" [ 559.399936] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.408092] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398590, 'name': CreateVM_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.470380] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.470980] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg c9b8af6fd1ea46f29ea5e238e5cf82e9 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 559.479780] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c9b8af6fd1ea46f29ea5e238e5cf82e9 [ 559.605447] env[62199]: DEBUG oslo_concurrency.lockutils [None req-f3ce8f11-8022-49ea-9bcb-38b9962823fd tempest-ServerDiagnosticsV248Test-736942293 tempest-ServerDiagnosticsV248Test-736942293-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.855154] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2609dea-4570-4e79-a828-0f895ff91dc3 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.863015] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4cea03e-c536-4d58-9ac9-f99d6c84bcd9 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.900267] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a6c0b3-2a4b-43d2-b138-208aa71f00ad {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.914080] env[62199]: DEBUG oslo_vmware.api [-] Task: {'id': task-398590, 'name': CreateVM_Task, 'duration_secs': 0.279288} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.915259] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b253c9d-c10a-41e5-beb2-9a00ace584b2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.922875] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Created VM on the ESX host {{(pid=62199) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 559.923330] env[62199]: DEBUG oslo_concurrency.lockutils [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 559.923483] env[62199]: DEBUG oslo_concurrency.lockutils [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.923807] env[62199]: DEBUG oslo_concurrency.lockutils [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 559.924478] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52853b68-023c-4b4e-bcca-59a7c0a4bf2f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.936938] env[62199]: DEBUG nova.compute.provider_tree [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.937424] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg 62d91ad169614f8584b3319d00ceade5 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 559.941402] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for the task: (returnval){ [ 559.941402] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]52371ea7-df2d-1a1b-9da9-4fef0dc027bf" [ 559.941402] env[62199]: _type = "Task" [ 559.941402] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.949503] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52371ea7-df2d-1a1b-9da9-4fef0dc027bf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.950251] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 62d91ad169614f8584b3319d00ceade5 [ 559.975524] env[62199]: DEBUG oslo_concurrency.lockutils [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Releasing lock "refresh_cache-24cb3e4a-dffd-4cb6-9ebd-151c3f11966f" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.976125] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62199) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 559.976985] env[62199]: DEBUG nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Deallocating network for instance {{(pid=62199) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 559.976985] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] deallocate_for_instance() {{(pid=62199) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 560.081019] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Instance cache missing network info. {{(pid=62199) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 560.081616] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg ad8fb6836464440f9730891fb79d48bb in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 560.094547] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ad8fb6836464440f9730891fb79d48bb [ 560.446949] env[62199]: DEBUG nova.scheduler.client.report [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Inventory has not changed for provider 715b0b11-2ed6-42ab-bb78-843b3217a820 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 126, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62199) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.446949] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg ad2c0786b8b4475b87609b1212492dc8 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 560.455530] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]52371ea7-df2d-1a1b-9da9-4fef0dc027bf, 'name': SearchDatastore_Task, 'duration_secs': 0.009256} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.455816] env[62199]: DEBUG oslo_concurrency.lockutils [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.456057] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Processing image 9fa830b9-8c21-4a7d-b7cb-64c85631deef {{(pid=62199) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 560.456294] env[62199]: DEBUG oslo_concurrency.lockutils [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.456432] env[62199]: DEBUG oslo_concurrency.lockutils [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.456597] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 560.457193] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ad2c0786b8b4475b87609b1212492dc8 [ 560.457567] env[62199]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-75a8162c-daf9-46de-a673-0cf3bb78563e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.469359] env[62199]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62199) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 560.469560] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62199) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 560.470432] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5900bb7-fed1-4c8c-ba3b-131fe1a8a0d2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.478136] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for the task: (returnval){ [ 560.478136] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]529e3e2d-6693-8ee7-05ff-6295da3f4fe2" [ 560.478136] env[62199]: _type = "Task" [ 560.478136] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.486261] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]529e3e2d-6693-8ee7-05ff-6295da3f4fe2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.583847] env[62199]: DEBUG nova.network.neutron [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Updating instance_info_cache with network_info: [] {{(pid=62199) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.584439] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg 06887bc90c2c41d983165825c1dab8aa in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 560.596071] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 06887bc90c2c41d983165825c1dab8aa [ 560.612631] env[62199]: DEBUG oslo_concurrency.lockutils [None req-cf86b038-819a-4f31-b5ce-ecdf52b5796f tempest-ServersV294TestFqdnHostnames-203601733 tempest-ServersV294TestFqdnHostnames-203601733-project-member] Acquiring lock "3a6e0c6a-cdcf-440a-a494-e12d0546d49e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.614693] env[62199]: DEBUG oslo_concurrency.lockutils [None req-cf86b038-819a-4f31-b5ce-ecdf52b5796f tempest-ServersV294TestFqdnHostnames-203601733 tempest-ServersV294TestFqdnHostnames-203601733-project-member] Lock "3a6e0c6a-cdcf-440a-a494-e12d0546d49e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.948410] env[62199]: DEBUG oslo_concurrency.lockutils [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.948985] env[62199]: DEBUG nova.compute.manager [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] [instance: 7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2] Start building networks asynchronously for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 560.951176] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg b30c67af416f4567a210863b514293f7 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 560.952212] env[62199]: DEBUG oslo_concurrency.lockutils [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.378s {{(pid=62199) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.953846] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 45735edffcc64c79b613f83644982d39 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 560.988739] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]529e3e2d-6693-8ee7-05ff-6295da3f4fe2, 'name': SearchDatastore_Task, 'duration_secs': 0.024928} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 560.989684] env[62199]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c82fe58-e432-4a74-a2c3-59f5da573989 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.995248] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for the task: (returnval){ [ 560.995248] env[62199]: value = "session[52e623c5-1ba1-f5e5-763c-53461b308d67]529239e8-b205-b5df-97fe-fd512ee39e67" [ 560.995248] env[62199]: _type = "Task" [ 560.995248] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.999499] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b30c67af416f4567a210863b514293f7 [ 561.004087] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]529239e8-b205-b5df-97fe-fd512ee39e67, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.006758] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 45735edffcc64c79b613f83644982d39 [ 561.090532] env[62199]: INFO nova.compute.manager [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] [instance: 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f] Took 1.11 seconds to deallocate network for instance. [ 561.090532] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg a96c7fad637a4d3d9c5d0304575d6713 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 561.133812] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a96c7fad637a4d3d9c5d0304575d6713 [ 561.465026] env[62199]: DEBUG nova.compute.utils [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Using /dev/sd instead of None {{(pid=62199) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.465841] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg 1a71de1f7bdf40adbf0bd250fc09a7da in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 561.469656] env[62199]: DEBUG nova.compute.manager [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] [instance: 7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2] Allocating IP information in the background. {{(pid=62199) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 561.469854] env[62199]: DEBUG nova.network.neutron [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] [instance: 7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2] allocate_for_instance() {{(pid=62199) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 561.478170] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1a71de1f7bdf40adbf0bd250fc09a7da [ 561.504907] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': session[52e623c5-1ba1-f5e5-763c-53461b308d67]529239e8-b205-b5df-97fe-fd512ee39e67, 'name': SearchDatastore_Task, 'duration_secs': 0.010506} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.507433] env[62199]: DEBUG oslo_concurrency.lockutils [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk" {{(pid=62199) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.507773] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore1] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 561.509165] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-02e8af5e-ef92-4ff0-b6ff-c1fd355ee710 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.516186] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for the task: (returnval){ [ 561.516186] env[62199]: value = "task-398592" [ 561.516186] env[62199]: _type = "Task" [ 561.516186] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.528632] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': task-398592, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.594668] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg d94c34b9c8824efa96716db713b229a6 in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 561.637801] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d94c34b9c8824efa96716db713b229a6 [ 561.826968] env[62199]: DEBUG nova.policy [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae8fde8ff47c43099cf864709df4024d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '47ddf80bc1da47979e34198e66478e58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62199) authorize /opt/stack/nova/nova/policy.py:203}} [ 561.853539] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-896f2bed-d39a-40c2-bd06-73749473659f {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.867374] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a62669-008e-43d0-8f41-4f7cbc333cee {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.907922] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d62ba1e-4bed-4042-befd-f8c60044d2a2 {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.920973] env[62199]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd051160-231c-42b0-b06c-1ed1e1ba281a {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.935344] env[62199]: DEBUG nova.compute.provider_tree [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Inventory has not changed in ProviderTree for provider: 715b0b11-2ed6-42ab-bb78-843b3217a820 {{(pid=62199) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 561.936121] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-ServerExternalEventsTest-1889348239 tempest-ServerExternalEventsTest-1889348239-project-member] Expecting reply to msg 06c58cd3452d4a13a3b83a5bea233afa in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 561.953485] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 06c58cd3452d4a13a3b83a5bea233afa [ 561.973355] env[62199]: DEBUG nova.compute.manager [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] [instance: 7d0094fa-4b0d-48c1-9ff7-e8b7478fd6e2] Start building block device mappings for instance. {{(pid=62199) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 561.975448] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-112ca4c1-dfa3-4f7a-bf4d-1377c6ce1a7f tempest-VolumesAssistedSnapshotsTest-281981774 tempest-VolumesAssistedSnapshotsTest-281981774-project-member] Expecting reply to msg 34e62c4de5df415aa9edd2c323f715da in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 562.023922] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 34e62c4de5df415aa9edd2c323f715da [ 562.029587] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': task-398592, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.461917} completed successfully. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 562.029857] env[62199]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/9fa830b9-8c21-4a7d-b7cb-64c85631deef/9fa830b9-8c21-4a7d-b7cb-64c85631deef.vmdk to [datastore1] 47b420a2-5eeb-4f14-9137-fb591a9dcbe6/47b420a2-5eeb-4f14-9137-fb591a9dcbe6.vmdk {{(pid=62199) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 562.030097] env[62199]: DEBUG nova.virt.vmwareapi.vmops [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] [instance: 47b420a2-5eeb-4f14-9137-fb591a9dcbe6] Extending root virtual disk to 1048576 {{(pid=62199) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 562.030411] env[62199]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-099216e7-e4bc-48fa-b36b-61b1fc01fa6e {{(pid=62199) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.036663] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Waiting for the task: (returnval){ [ 562.036663] env[62199]: value = "task-398593" [ 562.036663] env[62199]: _type = "Task" [ 562.036663] env[62199]: } to complete. {{(pid=62199) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 562.050309] env[62199]: DEBUG oslo_vmware.api [None req-b9327a89-7db6-4efc-a165-32a77a00accc tempest-ServersAdmin275Test-1064175422 tempest-ServersAdmin275Test-1064175422-project-admin] Task: {'id': task-398593, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62199) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.124326] env[62199]: INFO nova.scheduler.client.report [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Deleted allocations for instance 24cb3e4a-dffd-4cb6-9ebd-151c3f11966f [ 562.135031] env[62199]: INFO oslo_messaging._drivers.amqpdriver [None req-764c4ffe-4fc9-4bd7-8e16-25f7a8123154 tempest-DeleteServersAdminTestJSON-246729361 tempest-DeleteServersAdminTestJSON-246729361-project-member] Expecting reply to msg ff167af337274f7ea47b1b8d7d32dc9d in queue reply_2b94d7d99e464fa289aad055ffbc8a8b [ 562.158006] env[62199]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ff167af337274f7ea47b1b8d7d32dc9d [ 562.442444] env[62199]: DEBUG nova.scheduler.client.report [None req-9003a77a-c331-46b1-a573-7427c6663a9c tempest-Server