[ 479.257270] env[62599]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62599) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 479.257605] env[62599]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62599) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 479.257726] env[62599]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62599) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 479.258009] env[62599]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 479.342966] env[62599]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62599) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 479.352464] env[62599]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.009s {{(pid=62599) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 479.398206] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] Creating reply queue: reply_9dcc922abc674063ad7107323bf8efa6 [ 479.406427] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] Expecting reply to msg 0e523f65fd92411f9027b6a6ab08be7a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 479.419984] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0e523f65fd92411f9027b6a6ab08be7a [ 479.956521] env[62599]: INFO nova.virt.driver [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 480.027885] env[62599]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 480.028118] env[62599]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 480.028169] env[62599]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62599) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 483.172873] env[62599]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-2997e1ea-5303-45e1-9cd1-ef70a2c93064 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.189997] env[62599]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62599) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 483.190153] env[62599]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-106bdb7d-909f-4787-9abf-3eae6019f7ba {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.230801] env[62599]: INFO oslo_vmware.api [-] Successfully established new session; session ID is f98b9. [ 483.230976] env[62599]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.203s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 483.231541] env[62599]: INFO nova.virt.vmwareapi.driver [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] VMware vCenter version: 7.0.3 [ 483.235156] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb070331-c6f9-4622-aa7c-cebd094bb02d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.252579] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cc88615-fb16-45e5-8253-28b776e5d577 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.258537] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9188f7-7f5d-46bf-9499-55629d7f28d7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.265163] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82a32e4b-cd1e-495b-b2f2-fcfce28c2328 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.278279] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01957f0-73cd-4c62-a9da-e524e2e27fc9 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.284227] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab6b1469-de75-483e-9895-e44f2dd0a572 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.314746] env[62599]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-2a20591a-99c7-467c-b01c-f621851d9a7e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 483.320175] env[62599]: DEBUG nova.virt.vmwareapi.driver [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] Extension org.openstack.compute already exists. {{(pid=62599) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:224}} [ 483.322915] env[62599]: INFO nova.compute.provider_config [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 483.323599] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] Expecting reply to msg f75a96c6d653438d999075cd20b4a819 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 483.340630] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f75a96c6d653438d999075cd20b4a819 [ 483.826735] env[62599]: DEBUG nova.context [None req-e1582f14-760d-4c26-9513-ea7dab3aa4a4 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),07eb5136-6211-4ca1-bb79-3ae43500d00d(cell1) {{(pid=62599) load_cells /opt/stack/nova/nova/context.py:474}} [ 483.828950] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 483.829347] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 483.830040] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 483.830484] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Acquiring lock "07eb5136-6211-4ca1-bb79-3ae43500d00d" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 483.830677] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Lock "07eb5136-6211-4ca1-bb79-3ae43500d00d" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 483.831871] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Lock "07eb5136-6211-4ca1-bb79-3ae43500d00d" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 483.852622] env[62599]: INFO dbcounter [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Registered counter for database nova_cell0 [ 483.861261] env[62599]: INFO dbcounter [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Registered counter for database nova_cell1 [ 483.864424] env[62599]: DEBUG oslo_db.sqlalchemy.engines [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62599) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 483.865053] env[62599]: DEBUG oslo_db.sqlalchemy.engines [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62599) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 483.870010] env[62599]: ERROR nova.db.main.api [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 483.870010] env[62599]: result = function(*args, **kwargs) [ 483.870010] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 483.870010] env[62599]: return func(*args, **kwargs) [ 483.870010] env[62599]: File "/opt/stack/nova/nova/context.py", line 432, in gather_result [ 483.870010] env[62599]: result = fn(*args, **kwargs) [ 483.870010] env[62599]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 483.870010] env[62599]: return f(*args, **kwargs) [ 483.870010] env[62599]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 483.870010] env[62599]: return db.service_get_minimum_version(context, binaries) [ 483.870010] env[62599]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 483.870010] env[62599]: _check_db_access() [ 483.870010] env[62599]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 483.870010] env[62599]: stacktrace = ''.join(traceback.format_stack()) [ 483.870010] env[62599]: [ 483.871117] env[62599]: ERROR nova.db.main.api [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 483.871117] env[62599]: result = function(*args, **kwargs) [ 483.871117] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 483.871117] env[62599]: return func(*args, **kwargs) [ 483.871117] env[62599]: File "/opt/stack/nova/nova/context.py", line 432, in gather_result [ 483.871117] env[62599]: result = fn(*args, **kwargs) [ 483.871117] env[62599]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 483.871117] env[62599]: return f(*args, **kwargs) [ 483.871117] env[62599]: File "/opt/stack/nova/nova/objects/service.py", line 553, in _db_service_get_minimum_version [ 483.871117] env[62599]: return db.service_get_minimum_version(context, binaries) [ 483.871117] env[62599]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 483.871117] env[62599]: _check_db_access() [ 483.871117] env[62599]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 483.871117] env[62599]: stacktrace = ''.join(traceback.format_stack()) [ 483.871117] env[62599]: [ 483.871742] env[62599]: WARNING nova.objects.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 483.871742] env[62599]: WARNING nova.objects.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Failed to get minimum service version for cell 07eb5136-6211-4ca1-bb79-3ae43500d00d [ 483.872137] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Acquiring lock "singleton_lock" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 483.872325] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Acquired lock "singleton_lock" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 483.872573] env[62599]: DEBUG oslo_concurrency.lockutils [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Releasing lock "singleton_lock" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 483.872968] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Full set of CONF: {{(pid=62599) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 483.873057] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ******************************************************************************** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2600}} [ 483.873160] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] Configuration options gathered from: {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2601}} [ 483.873295] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2602}} [ 483.873487] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2603}} [ 483.873642] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ================================================================================ {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2605}} [ 483.873856] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] allow_resize_to_same_host = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.874028] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] arq_binding_timeout = 300 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.874158] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] backdoor_port = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.874286] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] backdoor_socket = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.874450] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] block_device_allocate_retries = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.874628] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] block_device_allocate_retries_interval = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.874814] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cert = self.pem {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.874982] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.875182] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute_monitors = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.875353] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] config_dir = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.875523] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] config_drive_format = iso9660 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.875658] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.875822] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] config_source = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.875988] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] console_host = devstack {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.876169] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] control_exchange = nova {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.876330] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cpu_allocation_ratio = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.876490] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] daemon = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.876655] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] debug = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.876814] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] default_access_ip_network_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.876980] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] default_availability_zone = nova {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.877139] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] default_ephemeral_format = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.877302] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] default_green_pool_size = 1000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.877646] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.877825] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] default_schedule_zone = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.877988] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] disk_allocation_ratio = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.878180] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] enable_new_services = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.878368] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] enabled_apis = ['osapi_compute'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.878535] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] enabled_ssl_apis = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.878697] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] flat_injected = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.878859] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] force_config_drive = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.879018] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] force_raw_images = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.879190] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] graceful_shutdown_timeout = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.879355] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] heal_instance_info_cache_interval = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.879572] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] host = cpu-1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.879836] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.880023] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.880197] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.880415] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.880586] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instance_build_timeout = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.880750] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instance_delete_interval = 300 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.880919] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instance_format = [instance: %(uuid)s] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.881103] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instance_name_template = instance-%08x {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.881285] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instance_usage_audit = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.881462] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instance_usage_audit_period = month {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.881634] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.881805] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.881973] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] internal_service_availability_zone = internal {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.882131] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] key = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.882292] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] live_migration_retry_count = 30 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.882459] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_color = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.882626] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_config_append = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.882795] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.882955] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_dir = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.883113] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.883244] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_options = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.883405] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_rotate_interval = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.883576] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_rotate_interval_type = days {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.883849] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] log_rotation_type = none {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.884017] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.884154] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.884331] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.884499] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.884630] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.884799] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] long_rpc_timeout = 1800 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.884962] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] max_concurrent_builds = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.885121] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] max_concurrent_live_migrations = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.885280] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] max_concurrent_snapshots = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.885441] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] max_local_block_devices = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.885601] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] max_logfile_count = 30 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.885761] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] max_logfile_size_mb = 200 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.885955] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] maximum_instance_delete_attempts = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.886124] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] metadata_listen = 0.0.0.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.886302] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] metadata_listen_port = 8775 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.886477] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] metadata_workers = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.886642] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] migrate_max_retries = -1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.886816] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] mkisofs_cmd = genisoimage {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.887028] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.887163] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] my_ip = 10.180.1.21 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.887332] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] network_allocate_retries = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.887612] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.887808] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.887980] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] osapi_compute_listen_port = 8774 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.888179] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] osapi_compute_unique_server_name_scope = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.888351] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] osapi_compute_workers = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.888517] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] password_length = 12 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.888680] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] periodic_enable = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.888847] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] periodic_fuzzy_delay = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.889041] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] pointer_model = usbtablet {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.889226] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] preallocate_images = none {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.889392] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] publish_errors = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.889526] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] pybasedir = /opt/stack/nova {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.889687] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ram_allocation_ratio = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.889949] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] rate_limit_burst = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.890130] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] rate_limit_except_level = CRITICAL {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.890295] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] rate_limit_interval = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.890458] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] reboot_timeout = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.890623] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] reclaim_instance_interval = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.890783] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] record = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.890955] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] reimage_timeout_per_gb = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.891123] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] report_interval = 120 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.891286] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] rescue_timeout = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.891447] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] reserved_host_cpus = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.891607] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] reserved_host_disk_mb = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.891770] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] reserved_host_memory_mb = 512 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.891930] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] reserved_huge_pages = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.892133] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] resize_confirm_window = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.892304] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] resize_fs_using_block_device = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.892465] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] resume_guests_state_on_host_boot = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.892637] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.892804] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] rpc_response_timeout = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.892966] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] run_external_periodic_tasks = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.893137] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] running_deleted_instance_action = reap {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.893303] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.893462] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] running_deleted_instance_timeout = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.893646] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler_instance_sync_interval = 120 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.893849] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_down_time = 720 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.894027] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] servicegroup_driver = db {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.894194] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] shelved_offload_time = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.894359] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] shelved_poll_interval = 3600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.894529] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] shutdown_timeout = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.894693] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] source_is_ipv6 = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.894858] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ssl_only = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.895166] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.895361] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] sync_power_state_interval = 600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.895532] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] sync_power_state_pool_size = 1000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.895706] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] syslog_log_facility = LOG_USER {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.895869] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] tempdir = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.896044] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] timeout_nbd = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.896220] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] transport_url = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.896384] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] update_resources_interval = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.896546] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] use_cow_images = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.896727] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] use_eventlog = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.896910] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] use_journal = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.897074] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] use_json = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.897235] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] use_rootwrap_daemon = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.897394] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] use_stderr = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.897631] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] use_syslog = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.897824] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vcpu_pin_set = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.898002] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plugging_is_fatal = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.898204] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plugging_timeout = 300 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.898380] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] virt_mkfs = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.898547] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] volume_usage_poll_interval = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.898711] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] watch_log_file = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.898884] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] web = /usr/share/spice-html5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2613}} [ 483.899070] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_concurrency.disable_process_locking = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.899361] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.899551] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.899735] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.900035] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.900233] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.900405] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.900590] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.auth_strategy = keystone {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.900762] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.compute_link_prefix = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.900938] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.901136] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.dhcp_domain = novalocal {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.901327] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.enable_instance_password = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.901498] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.glance_link_prefix = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.901666] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.901842] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.902005] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.instance_list_per_project_cells = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.902166] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.list_records_by_skipping_down_cells = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.902328] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.local_metadata_per_cell = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.902494] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.max_limit = 1000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.902668] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.metadata_cache_expiration = 15 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.902874] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.neutron_default_tenant_id = default {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.903057] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.response_validation = warn {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.903228] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.use_neutron_default_nets = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.903400] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.903564] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.903766] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.903955] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.904193] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.vendordata_dynamic_targets = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.904376] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.vendordata_jsonfile_path = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.904570] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.904760] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.backend = dogpile.cache.memcached {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.904933] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.backend_argument = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.905105] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.config_prefix = cache.oslo {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.905278] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.dead_timeout = 60.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.905445] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.debug_cache_backend = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.905613] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.enable_retry_client = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.905783] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.enable_socket_keepalive = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.905981] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.enabled = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.906152] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.enforce_fips_mode = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.906319] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.expiration_time = 600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.906485] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.hashclient_retry_attempts = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.906655] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.906821] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_dead_retry = 300 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.906983] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_password = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.907164] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.907353] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.907524] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_pool_maxsize = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.907770] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.907966] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_sasl_enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.908170] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.908344] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.908509] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.memcache_username = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.908679] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.proxies = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.908876] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.redis_db = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.909062] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.redis_password = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.909242] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.909423] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.909597] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.redis_server = localhost:6379 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.909765] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.redis_socket_timeout = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.909996] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.redis_username = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.910206] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.retry_attempts = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.910395] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.retry_delay = 0.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.910566] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.socket_keepalive_count = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.910730] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.socket_keepalive_idle = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.910894] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.socket_keepalive_interval = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.911051] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.tls_allowed_ciphers = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.911209] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.tls_cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.911363] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.tls_certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.911528] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.tls_enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.911687] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cache.tls_keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.911874] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.912094] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.auth_type = password {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.912270] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.912455] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.912622] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.912793] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.912960] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.cross_az_attach = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.913127] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.debug = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.913314] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.endpoint_template = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.913493] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.http_retries = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.913689] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.913867] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.914043] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.os_region_name = RegionOne {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.914212] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.914374] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cinder.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.914549] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.914740] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.cpu_dedicated_set = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.914911] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.cpu_shared_set = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.915083] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.image_type_exclude_list = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.915249] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.915416] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.915581] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.915745] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.915928] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.916136] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.resource_provider_association_refresh = 300 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.916310] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.916474] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.shutdown_retry_interval = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.916657] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.916839] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] conductor.workers = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.917017] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] console.allowed_origins = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.917178] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] console.ssl_ciphers = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.917347] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] console.ssl_minimum_version = default {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.917518] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] consoleauth.enforce_session_timeout = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.917713] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] consoleauth.token_ttl = 600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.917989] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.918167] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.918338] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.918502] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.connect_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.918666] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.connect_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.918827] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.endpoint_override = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.919010] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.919190] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.919357] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.max_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.919522] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.min_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.919686] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.region_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.919853] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.retriable_status_codes = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.920033] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.service_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.920216] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.service_type = accelerator {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.920383] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.920546] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.status_code_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.920730] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.status_code_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.920908] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.921093] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.921254] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] cyborg.version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.921434] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.backend = sqlalchemy {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.921606] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.connection = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.921778] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.connection_debug = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.921950] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.connection_parameters = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.922181] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.connection_recycle_time = 3600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.922367] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.connection_trace = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.922537] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.db_inc_retry_interval = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.922704] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.db_max_retries = 20 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.922871] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.db_max_retry_interval = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.923038] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.db_retry_interval = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.923203] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.max_overflow = 50 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.923368] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.max_pool_size = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.923531] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.max_retries = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.923758] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.923935] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.mysql_wsrep_sync_wait = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.924113] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.pool_timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.924283] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.retry_interval = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.924446] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.slave_connection = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.924612] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.sqlite_synchronous = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.924778] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] database.use_db_reconnect = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.924958] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.backend = sqlalchemy {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.925160] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.connection = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.925339] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.connection_debug = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.925514] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.connection_parameters = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.925684] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.connection_recycle_time = 3600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.925851] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.connection_trace = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.926016] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.db_inc_retry_interval = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.926182] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.db_max_retries = 20 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.926346] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.db_max_retry_interval = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.926510] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.db_retry_interval = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.926675] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.max_overflow = 50 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.926873] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.max_pool_size = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.927043] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.max_retries = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.927216] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.927378] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.927539] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.pool_timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.927706] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.retry_interval = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.927870] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.slave_connection = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.928172] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] api_database.sqlite_synchronous = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.928373] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] devices.enabled_mdev_types = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.928557] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.928734] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.928904] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ephemeral_storage_encryption.enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.929074] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.929248] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.api_servers = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.929417] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.929585] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.929770] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.929955] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.connect_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.930121] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.connect_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.930288] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.debug = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.930459] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.default_trusted_certificate_ids = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.930627] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.enable_certificate_validation = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.930791] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.enable_rbd_download = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.930951] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.endpoint_override = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.931141] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.931321] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.931490] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.max_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.931646] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.min_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.931812] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.num_retries = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.931983] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.rbd_ceph_conf = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.932165] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.rbd_connect_timeout = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.932337] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.rbd_pool = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.932507] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.rbd_user = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.932669] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.region_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.932859] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.retriable_status_codes = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.933034] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.service_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.933210] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.service_type = image {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.933375] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.933537] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.status_code_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.933730] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.status_code_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.933904] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.934090] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.934297] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.verify_glance_signatures = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.934464] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] glance.version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.934643] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] guestfs.debug = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.934808] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] healthcheck.cache_control = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.934976] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] healthcheck.debug = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.935141] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] healthcheck.ttl = 300 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.935303] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] healthcheck.uri = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.935472] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] mks.enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.935842] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.936064] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] image_cache.manager_interval = 2400 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.936243] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] image_cache.precache_concurrency = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.936418] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] image_cache.remove_unused_base_images = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.936592] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.936767] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.937014] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] image_cache.subdirectory_name = _base {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.937238] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.api_max_retries = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.937424] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.api_retry_interval = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.937592] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.937761] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.auth_type = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.937924] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.938172] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.938356] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.938528] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.conductor_group = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.938694] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.connect_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.938896] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.connect_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.939111] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.endpoint_override = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.939289] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.939454] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.939620] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.max_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.939783] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.min_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.939951] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.peer_list = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.940127] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.region_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.940323] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.retriable_status_codes = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.940501] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.serial_console_state_timeout = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.940666] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.service_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.940845] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.service_type = baremetal {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.941008] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.shard = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.941174] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.941336] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.status_code_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.941498] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.status_code_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.941680] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.941893] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.942080] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ironic.version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.942269] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.942449] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] key_manager.fixed_key = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.942636] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.942803] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.barbican_api_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.942965] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.barbican_endpoint = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.943140] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.barbican_endpoint_type = public {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.943329] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.barbican_region_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.943502] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.943694] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.943870] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.944050] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.944218] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.944387] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.number_of_retries = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.944554] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.retry_delay = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.944722] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.send_service_user_token = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.944888] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.945051] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.945216] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.verify_ssl = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.945378] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican.verify_ssl_path = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.945550] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.945716] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.auth_type = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.945879] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.946074] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.946251] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.946415] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.946576] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.946745] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.946973] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] barbican_service_user.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.947162] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.approle_role_id = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.947327] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.approle_secret_id = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.947499] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.kv_mountpoint = secret {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.947662] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.kv_path = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.947831] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.kv_version = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.947995] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.namespace = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.948250] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.root_token_id = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.948432] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.ssl_ca_crt_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.948608] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.timeout = 60.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.948776] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.use_ssl = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.948953] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.949155] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.949325] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.auth_type = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.949486] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.949647] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.949818] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.949980] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.connect_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.950142] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.connect_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.950300] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.endpoint_override = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.950463] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.950622] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.950782] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.max_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.950940] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.min_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.951123] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.region_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.951333] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.retriable_status_codes = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.951504] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.service_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.951825] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.service_type = identity {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.951825] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.951995] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.status_code_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.952201] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.status_code_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.952370] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.952558] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.952723] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] keystone.version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.952928] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.connection_uri = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.953092] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.cpu_mode = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.953328] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.953423] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.cpu_models = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.953627] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.cpu_power_governor_high = performance {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.953816] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.953947] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.cpu_power_management = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.954124] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.954294] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.device_detach_attempts = 8 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.954459] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.device_detach_timeout = 20 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.954629] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.disk_cachemodes = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.954792] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.disk_prefix = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.954962] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.enabled_perf_events = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.955168] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.file_backed_memory = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.955352] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.gid_maps = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.955516] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.hw_disk_discard = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.955677] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.hw_machine_type = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.955849] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.images_rbd_ceph_conf = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.956028] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.956203] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.956374] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.images_rbd_glance_store_name = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.956543] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.images_rbd_pool = rbd {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.956715] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.images_type = default {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.956878] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.images_volume_group = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.957042] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.inject_key = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.957206] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.inject_partition = -2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.957367] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.inject_password = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.957547] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.iscsi_iface = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.957686] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.iser_use_multipath = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.957853] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.958016] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.958209] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_downtime = 500 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.958477] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.958656] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.958827] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_inbound_addr = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.959030] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.959210] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.959374] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_scheme = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.959552] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_timeout_action = abort {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.959723] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_tunnelled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.959885] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_uri = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.961921] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.live_migration_with_native_tls = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.961921] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.max_queues = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.961921] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.961921] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.961921] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.nfs_mount_options = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.961921] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.961921] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.962387] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.962387] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.962387] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.962560] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.num_pcie_ports = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.962739] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.962919] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.pmem_namespaces = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.963086] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.quobyte_client_cfg = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.963398] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.963584] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.963786] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.963966] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.964661] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rbd_secret_uuid = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.964661] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rbd_user = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.964661] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.964661] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.964862] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rescue_image_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.964968] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rescue_kernel_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.965134] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rescue_ramdisk_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.965308] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.965473] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.rx_queue_size = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.965646] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.smbfs_mount_options = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.965936] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.966113] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.snapshot_compression = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.966275] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.snapshot_image_format = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.966496] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.966666] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.sparse_logical_volumes = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.966832] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.swtpm_enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.967002] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.swtpm_group = tss {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.967174] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.swtpm_user = tss {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.967344] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.sysinfo_serial = unique {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.967509] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.tb_cache_size = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.967670] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.tx_queue_size = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.967841] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.uid_maps = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.968054] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.use_virtio_for_bridges = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.968197] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.virt_type = kvm {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.968458] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.volume_clear = zero {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.968651] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.volume_clear_size = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.968829] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.volume_use_multipath = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.968994] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.vzstorage_cache_path = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.969171] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.969344] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.969514] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.969689] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.969971] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.970151] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.vzstorage_mount_user = stack {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.970320] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.970495] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.970669] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.auth_type = password {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.970834] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.971019] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.971199] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.971363] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.connect_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.971525] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.connect_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.971697] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.default_floating_pool = public {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.971860] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.endpoint_override = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.972038] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.extension_sync_interval = 600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.972209] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.http_retries = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.972372] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.972531] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.972691] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.max_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.972862] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.973024] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.min_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.973195] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.ovs_bridge = br-int {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.973361] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.physnets = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.973531] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.region_name = RegionOne {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.973724] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.retriable_status_codes = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.973894] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.service_metadata_proxy = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.974053] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.service_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.974220] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.service_type = network {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.974381] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.974542] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.status_code_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.974709] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.status_code_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.974866] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.975049] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.975213] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] neutron.version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.975385] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] notifications.bdms_in_notifications = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.975563] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] notifications.default_level = INFO {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.975738] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] notifications.notification_format = unversioned {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.975904] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] notifications.notify_on_state_change = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.976093] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.976274] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] pci.alias = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.976442] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] pci.device_spec = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.976609] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] pci.report_in_placement = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.976783] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.976956] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.auth_type = password {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.977123] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.977283] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.977441] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.977606] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.977770] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.connect_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.977932] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.connect_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.978090] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.default_domain_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.978249] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.default_domain_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.978405] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.domain_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.978661] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.domain_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.978838] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.endpoint_override = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.979006] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.979168] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.979326] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.max_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.979484] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.min_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.979654] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.password = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.979817] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.project_domain_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.979984] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.project_domain_name = Default {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.980168] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.project_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.980344] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.project_name = service {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.980514] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.region_name = RegionOne {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.980680] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.retriable_status_codes = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.980843] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.service_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.981013] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.service_type = placement {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.981176] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.981336] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.status_code_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.981501] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.status_code_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.981683] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.system_scope = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.981859] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.982021] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.trust_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.982182] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.user_domain_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.982351] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.user_domain_name = Default {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.982514] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.user_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.982690] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.username = nova {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.982875] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.983036] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] placement.version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.983217] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.cores = 20 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.983384] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.count_usage_from_placement = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.983561] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.983768] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.injected_file_content_bytes = 10240 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.983948] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.injected_file_path_length = 255 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.984139] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.injected_files = 5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.984311] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.instances = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.984480] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.key_pairs = 100 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.984647] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.metadata_items = 128 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.984819] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.ram = 51200 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.984982] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.recheck_quota = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.985153] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.server_group_members = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.985320] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] quota.server_groups = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.985488] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.985654] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.985821] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.image_metadata_prefilter = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.985985] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.986153] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.max_attempts = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.986320] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.max_placement_results = 1000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.986488] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.986653] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.986822] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.986996] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] scheduler.workers = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.987170] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.987343] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.987527] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.987700] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.987873] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.988051] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.988215] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.988402] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.988647] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.host_subset_size = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.988847] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.989014] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.989186] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.989357] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.isolated_hosts = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.989525] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.isolated_images = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.989692] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.989858] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.990024] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.990186] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.pci_in_placement = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.990350] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.990515] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.990679] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.990844] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.991009] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.991173] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.991335] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.track_instance_changes = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.991514] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.991713] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] metrics.required = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.991895] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] metrics.weight_multiplier = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.992099] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.992298] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] metrics.weight_setting = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.992621] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.992800] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] serial_console.enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.992979] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] serial_console.port_range = 10000:20000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.993152] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.993320] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.993490] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] serial_console.serialproxy_port = 6083 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.993687] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.993871] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.auth_type = password {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.994033] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.994193] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.994358] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.994520] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.994737] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.994860] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.send_service_user_token = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.995024] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.995199] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] service_user.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.995370] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.agent_enabled = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.995534] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.995854] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.996069] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.996250] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.html5proxy_port = 6082 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.996413] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.image_compression = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.996581] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.jpeg_compression = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.996744] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.playback_compression = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.996919] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.server_listen = 127.0.0.1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.997092] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.997254] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.streaming_mode = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.997412] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] spice.zlib_compression = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 483.999826] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] upgrade_levels.baseapi = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.000043] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] upgrade_levels.compute = auto {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.000223] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] upgrade_levels.conductor = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.000391] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] upgrade_levels.scheduler = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.000562] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.000726] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.000892] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.001051] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.001217] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.001378] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.001539] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.001757] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.001939] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vendordata_dynamic_auth.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.002118] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.api_retry_count = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.002282] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.ca_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.002457] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.002629] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.cluster_name = testcl1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.002801] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.connection_pool_size = 10 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.002963] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.console_delay_seconds = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.003136] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.datastore_regex = ^datastore.* {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.003348] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.003522] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.host_password = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.003735] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.host_port = 443 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.003917] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.host_username = administrator@vsphere.local {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.004104] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.insecure = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.004273] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.integration_bridge = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.004440] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.maximum_objects = 100 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.004621] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.pbm_default_policy = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.004800] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.pbm_enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.004964] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.pbm_wsdl_location = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.005135] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.005298] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.serial_port_proxy_uri = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.005456] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.serial_port_service_uri = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.005624] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.task_poll_interval = 0.5 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.005797] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.use_linked_clone = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.005975] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.vnc_keymap = en-us {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.006139] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.vnc_port = 5900 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.006303] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vmware.vnc_port_total = 10000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.006495] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.auth_schemes = ['none'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.006673] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.006975] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.007162] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.007335] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.novncproxy_port = 6080 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.007513] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.server_listen = 127.0.0.1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.007686] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.007849] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.vencrypt_ca_certs = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.008014] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.vencrypt_client_cert = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.008178] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vnc.vencrypt_client_key = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.008434] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.008530] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.disable_deep_image_inspection = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.008674] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.008837] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.008999] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.009161] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.disable_rootwrap = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.009322] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.enable_numa_live_migration = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.009484] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.009645] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.009828] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.009989] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.libvirt_disable_apic = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.010147] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.010309] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.010470] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.010631] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.010796] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.010956] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.011115] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.011274] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.011435] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.011613] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.011817] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.011993] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.client_socket_timeout = 900 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.012178] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.default_pool_size = 1000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.012347] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.keep_alive = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.012516] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.max_header_line = 16384 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.012683] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.012848] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.ssl_ca_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.013009] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.ssl_cert_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.013169] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.ssl_key_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.013333] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.tcp_keepidle = 600 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.013506] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.013705] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] zvm.ca_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.013903] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] zvm.cloud_connector_url = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.014202] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.014381] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] zvm.reachable_timeout = 300 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.014565] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.enforce_new_defaults = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.014772] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.enforce_scope = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.014959] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.policy_default_rule = default {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.015146] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.015321] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.policy_file = policy.yaml {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.015496] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.015659] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.015821] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.015978] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.016156] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.016325] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.016500] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.016682] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.connection_string = messaging:// {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.016852] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.enabled = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.017021] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.es_doc_type = notification {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.017186] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.es_scroll_size = 10000 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.017356] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.es_scroll_time = 2m {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.017522] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.filter_error_trace = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.017694] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.hmac_keys = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.017864] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.sentinel_service_name = mymaster {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.018031] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.socket_timeout = 0.1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.018194] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.trace_requests = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.018354] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler.trace_sqlalchemy = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.018530] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler_jaeger.process_tags = {} {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.018693] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler_jaeger.service_name_prefix = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.018861] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] profiler_otlp.service_name_prefix = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.019028] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] remote_debug.host = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.019188] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] remote_debug.port = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.019367] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.019530] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.019695] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.019884] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.020021] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.020184] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.020352] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.020516] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.020682] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.020856] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.021019] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.021192] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.021359] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.021531] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.021727] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.021905] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.022070] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.022246] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.022413] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.022576] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.022744] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.022911] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.023075] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.023244] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.023408] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.023568] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.023763] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.023933] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.024116] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.024286] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.ssl = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.024460] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.024650] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.024828] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.025003] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.025180] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.025337] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.025526] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.025695] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_notifications.retry = -1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.025912] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.026098] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.026271] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.auth_section = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.026436] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.auth_type = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.026598] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.cafile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.026761] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.certfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.026924] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.collect_timing = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.027085] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.connect_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.027245] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.connect_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.027403] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.endpoint_id = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.027561] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.endpoint_override = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.027722] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.insecure = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.027881] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.keyfile = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.028058] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.max_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.028225] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.min_version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.028384] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.region_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.028546] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.retriable_status_codes = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.028707] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.service_name = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.028868] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.service_type = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.029029] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.split_loggers = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.029188] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.status_code_retries = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.029350] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.status_code_retry_delay = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.029512] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.timeout = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.029676] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.valid_interfaces = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.029836] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_limit.version = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.029998] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_reports.file_event_handler = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.030164] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.030330] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] oslo_reports.log_dir = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.030606] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.030832] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.031006] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.031181] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.031345] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.031506] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.031681] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.031846] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_ovs_privileged.group = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.032016] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.032190] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.032357] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.032519] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] vif_plug_ovs_privileged.user = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.032692] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.032880] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.033056] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.033230] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.033401] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.033568] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.033824] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.034030] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.034218] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.034393] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_ovs.isolate_vif = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.034571] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.034778] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.034960] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.035136] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.035303] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_vif_ovs.per_port_bridge = False {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.035472] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] os_brick.lock_path = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.035643] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] privsep_osbrick.capabilities = [21] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.035806] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] privsep_osbrick.group = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.035966] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] privsep_osbrick.helper_command = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.036146] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.036313] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.036472] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] privsep_osbrick.user = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.036645] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.036812] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] nova_sys_admin.group = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.036970] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] nova_sys_admin.helper_command = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.037136] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.037301] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.037460] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] nova_sys_admin.user = None {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2620}} [ 484.037590] env[62599]: DEBUG oslo_service.service [None req-21f4b631-1c23-49f9-a676-49a93a7931f0 None None] ******************************************************************************** {{(pid=62599) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2624}} [ 484.038054] env[62599]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 484.038918] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 60efb806c7fb45b0a591d5b92faae1ce in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 484.047303] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 60efb806c7fb45b0a591d5b92faae1ce [ 484.541233] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Getting list of instances from cluster (obj){ [ 484.541233] env[62599]: value = "domain-c8" [ 484.541233] env[62599]: _type = "ClusterComputeResource" [ 484.541233] env[62599]: } {{(pid=62599) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 484.542387] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9347ae0d-9cad-41c9-9af8-897cc13195f0 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 484.551299] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Got total of 0 instances {{(pid=62599) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 484.551875] env[62599]: WARNING nova.virt.vmwareapi.driver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 484.552385] env[62599]: INFO nova.virt.node [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Generated node identity a361070e-e5e7-4d3d-83b7-c85d96b4b61c [ 484.552615] env[62599]: INFO nova.virt.node [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Wrote node identity a361070e-e5e7-4d3d-83b7-c85d96b4b61c to /opt/stack/data/n-cpu-1/compute_id [ 484.553034] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 4d7768dedcc94b648e4bdc06298b1584 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 484.564523] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4d7768dedcc94b648e4bdc06298b1584 [ 485.055805] env[62599]: WARNING nova.compute.manager [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Compute nodes ['a361070e-e5e7-4d3d-83b7-c85d96b4b61c'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 485.056575] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg aeab99068af34fc2b51f6c5962d408b6 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 485.081423] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg aeab99068af34fc2b51f6c5962d408b6 [ 485.558940] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 754ab6c9f2a244d3bc7ece5e9966cec3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 485.570700] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 754ab6c9f2a244d3bc7ece5e9966cec3 [ 486.061665] env[62599]: INFO nova.compute.manager [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 486.062183] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 6fb37a69c9834a81a48204af422de62a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 486.074014] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6fb37a69c9834a81a48204af422de62a [ 486.565361] env[62599]: INFO nova.healthcheck.manager [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] healthcheck endpoint is not enabled [ 486.565811] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 081a89333e894b3d824bd7bebe7e2666 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 486.578661] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 081a89333e894b3d824bd7bebe7e2666 [ 487.068107] env[62599]: WARNING nova.compute.manager [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 487.068379] env[62599]: DEBUG oslo_concurrency.lockutils [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 487.068613] env[62599]: DEBUG oslo_concurrency.lockutils [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 487.068767] env[62599]: DEBUG oslo_concurrency.lockutils [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 487.068922] env[62599]: DEBUG nova.compute.resource_tracker [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62599) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 487.069842] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2a9c9fc-8b4d-4281-85e4-a89e797be98d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.078072] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fda9754-5b2f-4b50-b914-5244ffa20e05 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.096746] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374d598f-2ba5-43cb-b4ef-ed39e50a6fd4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.104357] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-864377f1-e1df-411f-900c-eead6498df99 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 487.134237] env[62599]: DEBUG nova.compute.resource_tracker [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181773MB free_disk=129GB free_vcpus=48 pci_devices=None {{(pid=62599) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 487.134408] env[62599]: DEBUG oslo_concurrency.lockutils [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 487.134543] env[62599]: DEBUG oslo_concurrency.lockutils [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 487.134978] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 6c144c3ab62d40baae48abad110018ca in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 487.146322] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6c144c3ab62d40baae48abad110018ca [ 487.637405] env[62599]: WARNING nova.compute.resource_tracker [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] No compute node record for cpu-1:a361070e-e5e7-4d3d-83b7-c85d96b4b61c: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host a361070e-e5e7-4d3d-83b7-c85d96b4b61c could not be found. [ 487.638618] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 407c905fdfa84e2fb77d1114def3c156 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 487.650076] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 407c905fdfa84e2fb77d1114def3c156 [ 488.141853] env[62599]: INFO nova.compute.resource_tracker [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: a361070e-e5e7-4d3d-83b7-c85d96b4b61c [ 488.142357] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 16f11aa471694667b55ec5794c6d202b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 488.153378] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 16f11aa471694667b55ec5794c6d202b [ 488.645638] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 37f3c34d119e4183bb1bc9cbe1b638dc in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 488.665473] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 37f3c34d119e4183bb1bc9cbe1b638dc [ 489.147920] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 744f2b9257d9426c9ddf34b4c9309b02 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 489.169920] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 744f2b9257d9426c9ddf34b4c9309b02 [ 489.651031] env[62599]: DEBUG nova.compute.resource_tracker [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62599) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 489.651379] env[62599]: DEBUG nova.compute.resource_tracker [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62599) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 489.811663] env[62599]: INFO nova.scheduler.client.report [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] [req-57755ad3-66a1-452f-8670-badd0811d07c] Created resource provider record via placement API for resource provider with UUID a361070e-e5e7-4d3d-83b7-c85d96b4b61c and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 489.830562] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f132b897-0f6e-46e6-bf86-6d1d8422ba43 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.838295] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fea8589-38f4-481b-bdac-bf6d8e2db96d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.870409] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc4051b8-c9a7-4716-ad9d-3e5d7f4f2444 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.877755] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b83f72c-66f1-4e38-b44a-9822f5362ee4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 489.890729] env[62599]: DEBUG nova.compute.provider_tree [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 489.891294] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg ac2dc5755a08478b9d6afbb373db7c2f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 489.898349] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ac2dc5755a08478b9d6afbb373db7c2f [ 490.425612] env[62599]: DEBUG nova.scheduler.client.report [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Updated inventory for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 490.425916] env[62599]: DEBUG nova.compute.provider_tree [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Updating resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c generation from 0 to 1 during operation: update_inventory {{(pid=62599) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 490.426077] env[62599]: DEBUG nova.compute.provider_tree [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 490.482524] env[62599]: DEBUG nova.compute.provider_tree [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Updating resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c generation from 1 to 2 during operation: update_traits {{(pid=62599) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 490.482524] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Expecting reply to msg 930aefba49f247c6aab002b485572272 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 490.490454] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 930aefba49f247c6aab002b485572272 [ 490.981263] env[62599]: DEBUG nova.compute.resource_tracker [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62599) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 490.981263] env[62599]: DEBUG oslo_concurrency.lockutils [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.845s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 490.981263] env[62599]: DEBUG nova.service [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Creating RPC server for service compute {{(pid=62599) start /opt/stack/nova/nova/service.py:186}} [ 490.989740] env[62599]: INFO oslo.messaging._drivers.impl_rabbit [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Creating fanout queue: compute_fanout_7aee8e5f45cc4a06afc138884a370c92 [ 490.992792] env[62599]: DEBUG nova.service [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] Join ServiceGroup membership for this service compute {{(pid=62599) start /opt/stack/nova/nova/service.py:203}} [ 490.993198] env[62599]: DEBUG nova.servicegroup.drivers.db [None req-93e12685-f0a4-478e-b575-1b670cee4b12 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62599) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 495.995092] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 0d429968bd544674918bcf59169077fb in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 496.009913] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0d429968bd544674918bcf59169077fb [ 522.994819] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._sync_power_states {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 522.995463] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 214bd17e4f404ed0a3c24e1c8cec70ac in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 523.007278] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 214bd17e4f404ed0a3c24e1c8cec70ac [ 523.498086] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Getting list of instances from cluster (obj){ [ 523.498086] env[62599]: value = "domain-c8" [ 523.498086] env[62599]: _type = "ClusterComputeResource" [ 523.498086] env[62599]: } {{(pid=62599) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 523.499198] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22dd141d-58da-4231-ba11-a6f93895d1d4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.507941] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Got total of 0 instances {{(pid=62599) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 523.508208] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 523.508528] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Getting list of instances from cluster (obj){ [ 523.508528] env[62599]: value = "domain-c8" [ 523.508528] env[62599]: _type = "ClusterComputeResource" [ 523.508528] env[62599]: } {{(pid=62599) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 523.509382] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700e2dc6-a510-45b8-8aaa-5cc10186927d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.517816] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Got total of 0 instances {{(pid=62599) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 525.993339] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Acquiring lock "030814bf-2510-49a7-a19a-b2afac554287" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.993583] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Lock "030814bf-2510-49a7-a19a-b2afac554287" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.994112] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 4e38d2f183e9418880a4b91a49bc8c87 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 526.010574] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4e38d2f183e9418880a4b91a49bc8c87 [ 526.501368] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 526.501368] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 8c4505d65c714df18dc4cea4c8555a1c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 526.576315] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8c4505d65c714df18dc4cea4c8555a1c [ 527.034836] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.035127] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.038312] env[62599]: INFO nova.compute.claims [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 527.038500] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg c08b5c538abf4182a2a11471769cf88a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 527.129352] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c08b5c538abf4182a2a11471769cf88a [ 527.546122] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg a4a3192e5d0245fb84a89d575f09d7e3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 527.563244] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a4a3192e5d0245fb84a89d575f09d7e3 [ 528.089103] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6077df-7245-422a-8d0e-928ecfb3bf88 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.097460] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3262c216-6d35-456e-bfda-9d74a916d44f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.131325] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672ceb42-e239-4f41-9cab-58b9b6f9888e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.140179] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4f90896-c116-4223-9539-0245bcbf22f2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.159348] env[62599]: DEBUG nova.compute.provider_tree [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.160607] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg ff8845449c57448aa5ca05724df358de in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 528.167733] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ff8845449c57448aa5ca05724df358de [ 528.662868] env[62599]: DEBUG nova.scheduler.client.report [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 528.665314] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 45679caf890e41629248d8a935324640 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 528.680494] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 45679caf890e41629248d8a935324640 [ 529.168422] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.133s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.168929] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 529.170535] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 1da41bc82a3b41bca0968dc3b2049986 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 529.171405] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Acquiring lock "f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 529.171666] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Lock "f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.172151] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg cb93a3b40a4640d18a98d5b46c915004 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 529.188794] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg cb93a3b40a4640d18a98d5b46c915004 [ 529.230596] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1da41bc82a3b41bca0968dc3b2049986 [ 529.678110] env[62599]: DEBUG nova.compute.utils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 529.678110] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 8de3e8c4fb814a75813221a61889b98b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 529.678110] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 529.678110] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 529.686564] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 529.689386] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 4e1dee162ded43f6ad4446199426dc17 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 529.702096] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8de3e8c4fb814a75813221a61889b98b [ 529.743587] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4e1dee162ded43f6ad4446199426dc17 [ 529.941762] env[62599]: DEBUG nova.policy [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '144171d9504e4117a7bdaaaed11cd382', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '568a4d72ed444963af4be517501af3e1', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 530.187125] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 530.189005] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg f0b73b4315544c7ba03eb1cc8479ea1a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 530.213904] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.214163] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.217321] env[62599]: INFO nova.compute.claims [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 530.217479] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 9a85e28157e24293b2ec9860a209fabd in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 530.239490] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f0b73b4315544c7ba03eb1cc8479ea1a [ 530.266566] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9a85e28157e24293b2ec9860a209fabd [ 530.694466] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg bc2c7c1067e34926a0d9fe3e7b99a333 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 530.722745] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg d1cce831969c4dd99f29b743d03d7385 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 530.739039] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d1cce831969c4dd99f29b743d03d7385 [ 530.756601] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bc2c7c1067e34926a0d9fe3e7b99a333 [ 530.834457] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquiring lock "76004611-5b01-4b15-9411-c60d404bfd25" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 530.834710] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Lock "76004611-5b01-4b15-9411-c60d404bfd25" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 530.835234] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 06b2d04927334d2691c7eb7ae355aa9e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 530.854902] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 06b2d04927334d2691c7eb7ae355aa9e [ 531.073156] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Successfully created port: d5e9ef63-ba42-4db9-9fb1-2c37885b2085 {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 531.186848] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquiring lock "91d61cd2-a61f-4a8a-8868-3f1f5654e176" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.187053] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Lock "91d61cd2-a61f-4a8a-8868-3f1f5654e176" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.187516] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 3923aafa942e4cb3ace652a50f766c9d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 531.197575] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 531.202807] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3923aafa942e4cb3ace652a50f766c9d [ 531.231241] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 531.231557] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 531.231764] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 531.231956] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 531.232142] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 531.232291] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 531.232507] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 531.232666] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 531.233053] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 531.233221] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 531.233455] env[62599]: DEBUG nova.virt.hardware [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 531.234293] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c89002-5fd4-43d7-a4ee-9e50b34e4147 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.253889] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92bd94cb-76ea-40d7-9de4-54921f1f712b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.284166] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c133eeb-2b74-4686-9655-8ee6f8845fe2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.337446] env[62599]: DEBUG nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 531.339884] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg f8761662bd724fa69cbd3312bf9169fa in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 531.349659] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd0836ff-3d8f-42ca-81a6-ba28e6134e5e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.357370] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d55cabd3-3531-4415-877a-bb7eab8cbfe8 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.386844] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f8761662bd724fa69cbd3312bf9169fa [ 531.387812] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5405cc8f-0ebb-4b37-8913-ad405db749b4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.395790] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a7fbf63-9674-4da5-b0b6-c2b8d7ac4993 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.410295] env[62599]: DEBUG nova.compute.provider_tree [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 531.410873] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 91730d82475346d28050b4ca72a3e891 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 531.418155] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 91730d82475346d28050b4ca72a3e891 [ 531.489439] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Acquiring lock "a888fb44-7545-48be-9135-7eb690994cdf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.490149] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Lock "a888fb44-7545-48be-9135-7eb690994cdf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.490788] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 01ed300d6a454876a7f554192d1475fc in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 531.504157] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 01ed300d6a454876a7f554192d1475fc [ 531.689987] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 531.692335] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 9dbb6cb64bd94a39a4dec039181382a2 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 531.729387] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9dbb6cb64bd94a39a4dec039181382a2 [ 531.864247] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.913366] env[62599]: DEBUG nova.scheduler.client.report [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.915850] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 847f397b7932406bad03f9d0c3850372 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 531.926474] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 847f397b7932406bad03f9d0c3850372 [ 531.993023] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 531.995867] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg baf1e9719c44420c84775269be6e159d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 532.042451] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg baf1e9719c44420c84775269be6e159d [ 532.211988] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.418762] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.204s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.419490] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 532.421179] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 55680f0c2b3945e8a1f9f01efcda5f6f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 532.422953] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.561s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.425439] env[62599]: INFO nova.compute.claims [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 532.427280] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 5c5bac92a9bf4490a53b82fa31c9788d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 532.476140] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5c5bac92a9bf4490a53b82fa31c9788d [ 532.480290] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 55680f0c2b3945e8a1f9f01efcda5f6f [ 532.523196] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.927137] env[62599]: DEBUG nova.compute.utils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.928113] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 4bea04ace540413ba6d299e308414c5c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 532.928777] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 532.929177] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 532.933456] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 0e8c89b5af8c4c28bab92b689388ab0f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 532.946939] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4bea04ace540413ba6d299e308414c5c [ 532.947621] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0e8c89b5af8c4c28bab92b689388ab0f [ 532.948928] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 532.953005] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg bda55815209544cab5258aa91b1edf4c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 533.009640] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bda55815209544cab5258aa91b1edf4c [ 533.164234] env[62599]: DEBUG nova.policy [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4f7a21e2e38481bb05c4b31e1049aec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c8922d0635454c32a542aab62ee316c3', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 533.443385] env[62599]: ERROR nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. [ 533.443385] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 533.443385] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 533.443385] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 533.443385] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.443385] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 533.443385] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.443385] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 533.443385] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.443385] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 533.443385] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.443385] env[62599]: ERROR nova.compute.manager raise self.value [ 533.443385] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.443385] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 533.443385] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.443385] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 533.444071] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.444071] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 533.444071] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. [ 533.444071] env[62599]: ERROR nova.compute.manager [ 533.444071] env[62599]: Traceback (most recent call last): [ 533.444071] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 533.444071] env[62599]: listener.cb(fileno) [ 533.444071] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.444071] env[62599]: result = function(*args, **kwargs) [ 533.444071] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.444071] env[62599]: return func(*args, **kwargs) [ 533.444071] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 533.444071] env[62599]: raise e [ 533.444071] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 533.444071] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 533.444071] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.444071] env[62599]: created_port_ids = self._update_ports_for_instance( [ 533.444071] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.444071] env[62599]: with excutils.save_and_reraise_exception(): [ 533.444071] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.444071] env[62599]: self.force_reraise() [ 533.444071] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.444071] env[62599]: raise self.value [ 533.444071] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.444071] env[62599]: updated_port = self._update_port( [ 533.444071] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.444071] env[62599]: _ensure_no_port_binding_failure(port) [ 533.444071] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.444071] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 533.444896] env[62599]: nova.exception.PortBindingFailed: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. [ 533.444896] env[62599]: Removing descriptor: 15 [ 533.445401] env[62599]: ERROR nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] Traceback (most recent call last): [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] yield resources [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self.driver.spawn(context, instance, image_meta, [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] vm_ref = self.build_virtual_machine(instance, [ 533.445401] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] for vif in network_info: [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return self._sync_wrapper(fn, *args, **kwargs) [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self.wait() [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self[:] = self._gt.wait() [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return self._exit_event.wait() [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.445755] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] result = hub.switch() [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return self.greenlet.switch() [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] result = function(*args, **kwargs) [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return func(*args, **kwargs) [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] raise e [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] nwinfo = self.network_api.allocate_for_instance( [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] created_port_ids = self._update_ports_for_instance( [ 533.446146] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] with excutils.save_and_reraise_exception(): [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self.force_reraise() [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] raise self.value [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] updated_port = self._update_port( [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] _ensure_no_port_binding_failure(port) [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] raise exception.PortBindingFailed(port_id=port['id']) [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] nova.exception.PortBindingFailed: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. [ 533.446641] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] [ 533.447277] env[62599]: INFO nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Terminating instance [ 533.453167] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Acquiring lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.453167] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Acquired lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.453167] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 533.453167] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg cd4b6f453f38456fb9abed1716801c88 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 533.458529] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 61c1daf17ccf4518a030fefac999cdac in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 533.464591] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg cd4b6f453f38456fb9abed1716801c88 [ 533.506768] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 61c1daf17ccf4518a030fefac999cdac [ 533.540030] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e191d3f0-1747-4a6c-869a-20abc8412dc4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.548821] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8670e827-e757-4033-877e-74894c46c7bb {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.585819] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8cbef95-c5f5-45a7-bd03-d1abf10bbdab {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.595200] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5859f7f3-1362-4105-94a9-230a0481ec8e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.608064] env[62599]: DEBUG nova.compute.provider_tree [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 533.608580] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 4d10147415ae4c4591d8b9a500dfffeb in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 533.618151] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4d10147415ae4c4591d8b9a500dfffeb [ 533.867813] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Successfully created port: 86ad4574-224f-403d-a71f-544dee05e869 {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.961472] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 533.987205] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 533.987572] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 533.987638] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 533.987856] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 533.987971] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 533.988077] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 533.988294] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 533.988442] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 533.988645] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 533.988750] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 533.988922] env[62599]: DEBUG nova.virt.hardware [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 533.989817] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ba88fcf-78cc-4fdd-9a98-f8eb68c4ac56 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.001642] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a590e5df-2934-4bb6-9def-1b68ef89c294 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.112079] env[62599]: DEBUG nova.scheduler.client.report [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 534.115882] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 4a444568baac4a069010c556b0b37d14 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 534.133767] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4a444568baac4a069010c556b0b37d14 [ 534.145910] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 534.405060] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.405578] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 755ca0cf65f84ac5a52d1e21e2719dd3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 534.416742] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 755ca0cf65f84ac5a52d1e21e2719dd3 [ 534.619389] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.196s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 534.619925] env[62599]: DEBUG nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 534.621704] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 5412bb17c0a7468fbb53a2bc80125da4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 534.623031] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.411s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 534.624352] env[62599]: INFO nova.compute.claims [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 534.625927] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 0613b6622e6f4adab95c8676b00f9193 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 534.680798] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0613b6622e6f4adab95c8676b00f9193 [ 534.683328] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5412bb17c0a7468fbb53a2bc80125da4 [ 534.817322] env[62599]: DEBUG nova.compute.manager [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Received event network-changed-d5e9ef63-ba42-4db9-9fb1-2c37885b2085 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 534.817494] env[62599]: DEBUG nova.compute.manager [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Refreshing instance network info cache due to event network-changed-d5e9ef63-ba42-4db9-9fb1-2c37885b2085. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 534.817719] env[62599]: DEBUG oslo_concurrency.lockutils [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] Acquiring lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.912361] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Releasing lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.912834] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 534.913128] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 534.913333] env[62599]: DEBUG oslo_concurrency.lockutils [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] Acquired lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.913575] env[62599]: DEBUG nova.network.neutron [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Refreshing network info cache for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085 {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 534.914018] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] Expecting reply to msg 6e15e40e7da14a849c8349308023e90d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 534.916016] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d0aabb5-05dd-4b2f-94f6-4da3f8b3cdfd {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.925501] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-014ead0b-540f-4ceb-bc65-e91d3f1489e7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.936409] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6e15e40e7da14a849c8349308023e90d [ 534.947945] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 030814bf-2510-49a7-a19a-b2afac554287 could not be found. [ 534.948218] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 534.948611] env[62599]: INFO nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Took 0.04 seconds to destroy the instance on the hypervisor. [ 534.948860] env[62599]: DEBUG oslo.service.loopingcall [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 534.949338] env[62599]: DEBUG nova.compute.manager [-] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 534.949511] env[62599]: DEBUG nova.network.neutron [-] [instance: 030814bf-2510-49a7-a19a-b2afac554287] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 535.017922] env[62599]: DEBUG nova.network.neutron [-] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.018458] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg d353d277036f451780bfafdd85ef65b0 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 535.030987] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d353d277036f451780bfafdd85ef65b0 [ 535.132455] env[62599]: DEBUG nova.compute.utils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 535.132455] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 559409a4fbe94883920a246b49df3e0c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 535.132455] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 1f92106c1e2e4ecc8797afee195ab741 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 535.137271] env[62599]: DEBUG nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Not allocating networking since 'none' was specified. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1979}} [ 535.150381] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1f92106c1e2e4ecc8797afee195ab741 [ 535.150938] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 559409a4fbe94883920a246b49df3e0c [ 535.470130] env[62599]: DEBUG nova.network.neutron [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 535.520475] env[62599]: DEBUG nova.network.neutron [-] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.520972] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 25e14eaa5b534998bcfe1e78cd84aaba in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 535.536900] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 25e14eaa5b534998bcfe1e78cd84aaba [ 535.633972] env[62599]: DEBUG nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 535.635680] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 2875422f9eae42119cb3d44233c93224 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 535.689771] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2875422f9eae42119cb3d44233c93224 [ 535.724714] env[62599]: DEBUG nova.network.neutron [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.724714] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] Expecting reply to msg 125eb4a472d74b1cb8b2951b8df62c71 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 535.740177] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 125eb4a472d74b1cb8b2951b8df62c71 [ 535.756949] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bf41877-d3c9-49ff-8fd2-6f09cae35f03 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.767098] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16f90031-cd59-4dd1-95d7-183c9a3f3b1b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.813113] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49a484cb-65da-4acf-b7d6-ad4c1687850c {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.821945] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7da6ee3-b82f-4e5d-8b8f-c939cd195963 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.838920] env[62599]: DEBUG nova.compute.provider_tree [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 535.839602] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg f224791ee5b74aba90c926e056957d1d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 535.851926] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f224791ee5b74aba90c926e056957d1d [ 536.024424] env[62599]: INFO nova.compute.manager [-] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Took 1.07 seconds to deallocate network for instance. [ 536.029643] env[62599]: DEBUG nova.compute.claims [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 536.029862] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.034113] env[62599]: ERROR nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. [ 536.034113] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.034113] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 536.034113] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.034113] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.034113] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.034113] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.034113] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.034113] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.034113] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 536.034113] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.034113] env[62599]: ERROR nova.compute.manager raise self.value [ 536.034113] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.034113] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.034113] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.034113] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.034522] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.034522] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.034522] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. [ 536.034522] env[62599]: ERROR nova.compute.manager [ 536.034522] env[62599]: Traceback (most recent call last): [ 536.034522] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.034522] env[62599]: listener.cb(fileno) [ 536.034522] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.034522] env[62599]: result = function(*args, **kwargs) [ 536.034522] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.034522] env[62599]: return func(*args, **kwargs) [ 536.034522] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 536.034522] env[62599]: raise e [ 536.034522] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 536.034522] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 536.034522] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.034522] env[62599]: created_port_ids = self._update_ports_for_instance( [ 536.034522] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.034522] env[62599]: with excutils.save_and_reraise_exception(): [ 536.034522] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.034522] env[62599]: self.force_reraise() [ 536.034522] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.034522] env[62599]: raise self.value [ 536.034522] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.034522] env[62599]: updated_port = self._update_port( [ 536.034522] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.034522] env[62599]: _ensure_no_port_binding_failure(port) [ 536.034522] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.034522] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.035393] env[62599]: nova.exception.PortBindingFailed: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. [ 536.035393] env[62599]: Removing descriptor: 16 [ 536.035393] env[62599]: ERROR nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Traceback (most recent call last): [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] yield resources [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self.driver.spawn(context, instance, image_meta, [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.035393] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] vm_ref = self.build_virtual_machine(instance, [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] for vif in network_info: [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return self._sync_wrapper(fn, *args, **kwargs) [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self.wait() [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self[:] = self._gt.wait() [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return self._exit_event.wait() [ 536.035722] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] result = hub.switch() [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return self.greenlet.switch() [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] result = function(*args, **kwargs) [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return func(*args, **kwargs) [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] raise e [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] nwinfo = self.network_api.allocate_for_instance( [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 536.036112] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] created_port_ids = self._update_ports_for_instance( [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] with excutils.save_and_reraise_exception(): [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self.force_reraise() [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] raise self.value [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] updated_port = self._update_port( [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] _ensure_no_port_binding_failure(port) [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.036456] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] raise exception.PortBindingFailed(port_id=port['id']) [ 536.036831] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] nova.exception.PortBindingFailed: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. [ 536.036831] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] [ 536.036831] env[62599]: INFO nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Terminating instance [ 536.037458] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Acquiring lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.037967] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Acquired lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.038136] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 536.038555] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg bc53c4b957ef4f93acaf974677fb6fe9 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 536.045987] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bc53c4b957ef4f93acaf974677fb6fe9 [ 536.146309] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 951c336251874358b45002a6a53a7d2b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 536.202911] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 951c336251874358b45002a6a53a7d2b [ 536.228556] env[62599]: DEBUG oslo_concurrency.lockutils [req-c16ffe5b-2f8f-4608-8ee3-cbcaf59d666d req-4f2daf02-8b80-4a69-bf43-3d566fd8bbc6 service nova] Releasing lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 536.347404] env[62599]: DEBUG nova.scheduler.client.report [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 536.347404] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 9dc5402f4728498abc4a12367d7b6589 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 536.365865] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9dc5402f4728498abc4a12367d7b6589 [ 536.580688] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 536.648141] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.648832] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 7650d8100344403f95ffbe6b9ed28319 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 536.654613] env[62599]: DEBUG nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 536.660601] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7650d8100344403f95ffbe6b9ed28319 [ 536.680760] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 536.680998] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 536.681153] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 536.681333] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 536.681501] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 536.681667] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 536.681873] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 536.682027] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 536.682189] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 536.682344] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 536.682519] env[62599]: DEBUG nova.virt.hardware [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 536.683871] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84cf9569-aaf7-4184-a527-a191c301864e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.692593] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95fa1d20-8b83-4ac4-8089-23db5529ce53 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.709657] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Instance VIF info [] {{(pid=62599) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 536.720764] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62599) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 536.721129] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-744a8662-2342-45ea-865e-fc9d0017685a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.733448] env[62599]: INFO nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Created folder: OpenStack in parent group-v4. [ 536.733715] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Creating folder: Project (2d061a55384a4f098d6e8432fdb39b76). Parent ref: group-v107225. {{(pid=62599) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 536.734109] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0bd933f0-327a-457e-a760-a81c6db1fc9d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.744299] env[62599]: INFO nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Created folder: Project (2d061a55384a4f098d6e8432fdb39b76) in parent group-v107225. [ 536.744541] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Creating folder: Instances. Parent ref: group-v107226. {{(pid=62599) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 536.744796] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d526e8be-6a2f-40d2-a638-abde8a41f63f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.753553] env[62599]: INFO nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Created folder: Instances in parent group-v107226. [ 536.753833] env[62599]: DEBUG oslo.service.loopingcall [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 536.754020] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Creating VM on the ESX host {{(pid=62599) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 536.754353] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-af74e808-f3e0-417a-a97d-6bede592bebe {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.772272] env[62599]: DEBUG nova.compute.manager [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Received event network-changed-86ad4574-224f-403d-a71f-544dee05e869 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 536.772465] env[62599]: DEBUG nova.compute.manager [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Refreshing instance network info cache due to event network-changed-86ad4574-224f-403d-a71f-544dee05e869. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 536.772679] env[62599]: DEBUG oslo_concurrency.lockutils [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] Acquiring lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.776039] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Acquiring lock "b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.776039] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Lock "b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.776519] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 2d368abb2d7a46ba94570d284157804a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 536.785504] env[62599]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 536.785504] env[62599]: value = "task-394391" [ 536.785504] env[62599]: _type = "Task" [ 536.785504] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 536.790395] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2d368abb2d7a46ba94570d284157804a [ 536.795926] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394391, 'name': CreateVM_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 536.851150] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 536.852177] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 536.855020] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg d80256ae8daf44abb5c8a0e8bc3ea2e7 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 536.856572] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.335s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.862104] env[62599]: INFO nova.compute.claims [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 536.862292] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 62e4b61490db488baf35f9849a58a24f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 536.921816] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d80256ae8daf44abb5c8a0e8bc3ea2e7 [ 536.939157] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 62e4b61490db488baf35f9849a58a24f [ 537.154193] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Releasing lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.154193] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 537.154193] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 537.154193] env[62599]: DEBUG oslo_concurrency.lockutils [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] Acquired lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.154193] env[62599]: DEBUG nova.network.neutron [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Refreshing network info cache for port 86ad4574-224f-403d-a71f-544dee05e869 {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 537.156372] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] Expecting reply to msg 2c90af7a1c36497caf440bb6ce51099d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.156372] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dd9b38d9-c9fb-48f6-bb6e-825fd1c1ec4e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.168814] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22553b57-0a0a-4f44-892b-4b9b15c66b39 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.177032] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2c90af7a1c36497caf440bb6ce51099d [ 537.189281] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a could not be found. [ 537.189508] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 537.189718] env[62599]: INFO nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 537.189960] env[62599]: DEBUG oslo.service.loopingcall [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 537.190149] env[62599]: DEBUG nova.compute.manager [-] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 537.190241] env[62599]: DEBUG nova.network.neutron [-] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 537.235815] env[62599]: DEBUG nova.network.neutron [-] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.236445] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 5403717dba0b46d385a39951fa4254d4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.243531] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5403717dba0b46d385a39951fa4254d4 [ 537.278101] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 537.279997] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg c9ff29c18678469f98f197384a8cdf3f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.309183] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394391, 'name': CreateVM_Task, 'duration_secs': 0.350416} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 537.309183] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Created VM on the ESX host {{(pid=62599) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 537.312872] env[62599]: DEBUG oslo_vmware.service [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd3717fd-d939-4f88-a98c-b480ca213555 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.316322] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.316484] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.317104] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 537.317343] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a79c3fa-c463-4bd8-9f08-8bcb5e514ca5 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.323059] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 537.323059] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]52a5910d-82c2-0179-64d5-27207666d196" [ 537.323059] env[62599]: _type = "Task" [ 537.323059] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.329553] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c9ff29c18678469f98f197384a8cdf3f [ 537.340378] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]52a5910d-82c2-0179-64d5-27207666d196, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 537.365110] env[62599]: DEBUG nova.compute.utils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 537.365744] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 0dd02869c9bb4f08aa94ecdc1f6de1ac in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.368598] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 562c9a57e0b345e28df23ebf5addb159 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.370104] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 537.370522] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 537.379214] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 562c9a57e0b345e28df23ebf5addb159 [ 537.381355] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0dd02869c9bb4f08aa94ecdc1f6de1ac [ 537.619454] env[62599]: DEBUG nova.compute.manager [req-7600aac7-a815-4d5a-afe6-6dc3466bdb6e req-8a287694-96d6-4d81-a60c-28b0feb28009 service nova] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Received event network-vif-deleted-d5e9ef63-ba42-4db9-9fb1-2c37885b2085 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 537.704987] env[62599]: DEBUG nova.network.neutron [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 537.734972] env[62599]: DEBUG nova.policy [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bf01597064b84442b1f627f6a067deab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '925cea7d758d4245ac1770c576ef82ed', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 537.739088] env[62599]: DEBUG nova.network.neutron [-] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.739524] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 4f01a608c73d4769a6b045e0c6d25eef in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.747922] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4f01a608c73d4769a6b045e0c6d25eef [ 537.815287] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.836559] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.836832] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Processing image 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 537.837057] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.837194] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquired lock "[datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.837606] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 537.838003] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2242d5bf-54eb-4dc5-aff5-da42f536f1a9 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.855524] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 537.855701] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62599) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 537.856507] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11212c48-504d-4285-b8f5-0d6c94a6a55e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.864148] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90f93a66-57c7-48ec-9ba7-621d6ae2f85b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.873406] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 537.873406] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]52db93b6-7e4e-3b97-e241-4a3f4acb6bd6" [ 537.873406] env[62599]: _type = "Task" [ 537.873406] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 537.874424] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 537.876973] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 65e70868497742ffb3a13790432fa34f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.896887] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Preparing fetch location {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 537.898119] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Creating directory with path [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 537.898451] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac78a870-b4ef-49f0-a75a-0d34e96440a0 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.911122] env[62599]: DEBUG nova.network.neutron [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.912123] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] Expecting reply to msg ba79a896a3304fbfb8b3feb8bb01c9cd in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 537.931146] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ba79a896a3304fbfb8b3feb8bb01c9cd [ 537.933916] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Created directory with path [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 537.934235] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Fetch image to [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 537.934507] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Downloading image file data 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk on the data store datastore1 {{(pid=62599) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 537.935725] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce338b50-5bea-45ec-b4b6-ab655780c513 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.953075] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6170f174-117e-4e3f-819c-f52246bf8784 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.973858] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 65e70868497742ffb3a13790432fa34f [ 537.993148] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d04635-0c34-49f3-b18b-87c1ebe57d68 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.031594] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2741f113-37ff-493e-a0ff-468543940172 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.034799] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4a6485-09fd-4cb7-9436-9340bc6d4aec {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.043251] env[62599]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-e1c8854b-c858-4d5e-937f-78cb3342bf57 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.053794] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-918fb401-5549-4691-887a-518649adc45b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.088397] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a326579-34c0-492c-87c8-24783eb25b5f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.091258] env[62599]: DEBUG nova.virt.vmwareapi.images [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Downloading image file data 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to the data store datastore1 {{(pid=62599) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 538.099844] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d913c3a5-e76b-4238-9c00-74bb8ab0701c {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.115892] env[62599]: DEBUG nova.compute.provider_tree [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.115892] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 6d9895eba4b144a6ab1660dd8c0f6f94 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 538.125994] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6d9895eba4b144a6ab1660dd8c0f6f94 [ 538.127416] env[62599]: DEBUG nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 538.129673] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 5b7f459647a14349a471069da8a62e54 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 538.142604] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5b7f459647a14349a471069da8a62e54 [ 538.166512] env[62599]: DEBUG oslo_vmware.rw_handles [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62599) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 538.241105] env[62599]: INFO nova.compute.manager [-] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Took 1.05 seconds to deallocate network for instance. [ 538.243517] env[62599]: DEBUG nova.compute.claims [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 538.243716] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.394267] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 746f5f4cb4c149e89bc166e3592bb3d5 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 538.401029] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Successfully created port: 98a0d6aa-fdaf-42ed-9049-fec66cbc3237 {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 538.416073] env[62599]: DEBUG oslo_concurrency.lockutils [req-f8ba4ca0-89ad-4508-a81d-aebfb32c8385 req-0296c0de-358c-4fe6-8fed-21e2b31fcbbc service nova] Releasing lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.456781] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 746f5f4cb4c149e89bc166e3592bb3d5 [ 538.633505] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.777s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 538.634865] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 538.636208] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 0f4bf8970af845f198bda7638fc86135 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 538.641753] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 2.610s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.643695] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 1f1c5cbe1ab64e3da424c1106e8f3982 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 538.718524] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0f4bf8970af845f198bda7638fc86135 [ 538.723987] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1f1c5cbe1ab64e3da424c1106e8f3982 [ 538.866993] env[62599]: DEBUG oslo_vmware.rw_handles [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Completed reading data from the image iterator. {{(pid=62599) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 538.867164] env[62599]: DEBUG oslo_vmware.rw_handles [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62599) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 538.908753] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 538.925285] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquiring lock "66c56e34-39ee-41a3-94dc-6fd26a2cd726" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.925507] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Lock "66c56e34-39ee-41a3-94dc-6fd26a2cd726" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.926015] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 52e10b4f1e0e4635b3a8998751fe817f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 538.945823] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 538.946077] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 538.946255] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 538.946470] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 538.946631] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 538.948155] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 538.948155] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 538.948155] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 538.948155] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 538.948155] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 538.948382] env[62599]: DEBUG nova.virt.hardware [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 538.948570] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-923c6f8f-5abc-4fe0-9b40-967d0de29a8b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.951522] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 52e10b4f1e0e4635b3a8998751fe817f [ 538.960269] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16f27d3-62b3-4060-8161-9e29160054a8 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.006615] env[62599]: DEBUG nova.virt.vmwareapi.images [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Downloaded image file data 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk on the data store datastore1 {{(pid=62599) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 539.008563] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Caching image {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 539.008846] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Copying Virtual Disk [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk to [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 539.009122] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ea6a7d16-b8a3-4b58-a945-f38f8e526f36 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.016772] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 539.016772] env[62599]: value = "task-394392" [ 539.016772] env[62599]: _type = "Task" [ 539.016772] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 539.025359] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394392, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.146503] env[62599]: DEBUG nova.compute.utils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 539.147246] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 56220bbc69dc4b9fb375ed95d2424d0f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 539.148280] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 539.148497] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 539.170257] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 56220bbc69dc4b9fb375ed95d2424d0f [ 539.271035] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fad87f08-6ae0-47dd-8f03-0b9ad67d5e50 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.280247] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff110907-a18b-4913-b2ee-5f0e561ed4ef {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.316821] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6583dbd-8d57-468d-b73d-52a6978b07fc {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.327796] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-682e6335-4a25-4040-85a0-ba5c1d2367b0 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.349615] env[62599]: DEBUG nova.compute.provider_tree [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 539.350364] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 10e0e9a77d8f4a6fa9ba6764f4eb66c8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 539.361447] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 10e0e9a77d8f4a6fa9ba6764f4eb66c8 [ 539.406609] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.406939] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.407127] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Starting heal instance info cache {{(pid=62599) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9931}} [ 539.407288] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Rebuilding the list of instances to heal {{(pid=62599) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9935}} [ 539.408191] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg d4f79e33358c4c2fbfded946aa9625b9 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 539.428154] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d4f79e33358c4c2fbfded946aa9625b9 [ 539.428761] env[62599]: DEBUG nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 539.431389] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 604dc3f924ed433ba9ad3f07591274cf in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 539.464950] env[62599]: DEBUG nova.policy [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '35dc23702ab34c0e96d13f0780d48683', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '709305fc477440b68f5f5f788afe0281', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 539.491034] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 604dc3f924ed433ba9ad3f07591274cf [ 539.528276] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394392, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 539.651863] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 539.654029] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 7e9ec65889ae4f60bc21ca7aaa901dd8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 539.733336] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7e9ec65889ae4f60bc21ca7aaa901dd8 [ 539.853809] env[62599]: DEBUG nova.scheduler.client.report [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.856280] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 620a0c92f0a5492aaf9be049d65d4f8b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 539.872149] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 620a0c92f0a5492aaf9be049d65d4f8b [ 539.911102] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Skipping network cache update for instance because it is Building. {{(pid=62599) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9944}} [ 539.911275] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Skipping network cache update for instance because it is Building. {{(pid=62599) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9944}} [ 539.911421] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Skipping network cache update for instance because it is Building. {{(pid=62599) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9944}} [ 539.911548] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Skipping network cache update for instance because it is Building. {{(pid=62599) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9944}} [ 539.911670] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Didn't find any instances for network info cache update. {{(pid=62599) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10017}} [ 539.912122] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.912357] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.912553] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.912791] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.913062] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.913186] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.913349] env[62599]: DEBUG nova.compute.manager [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62599) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10551}} [ 539.913498] env[62599]: DEBUG oslo_service.periodic_task [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Running periodic task ComputeManager.update_available_resource {{(pid=62599) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 539.913928] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 6c82da232ca4466a8aa7e0d9478749cc in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 539.926742] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6c82da232ca4466a8aa7e0d9478749cc [ 539.956294] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.027384] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394392, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.705525} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.027636] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Copied Virtual Disk [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk to [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 540.027996] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Deleting the datastore file [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 540.028314] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ee3fce7c-3aac-4a98-808c-8e91b3a0dbbc {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.035286] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 540.035286] env[62599]: value = "task-394393" [ 540.035286] env[62599]: _type = "Task" [ 540.035286] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.043644] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394393, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.072469] env[62599]: DEBUG nova.compute.manager [req-cfa5f31a-8dc0-4404-a314-3fd041b9b32d req-ec35d7c8-f771-4ce9-91cb-9427caa9dcfa service nova] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Received event network-vif-deleted-86ad4574-224f-403d-a71f-544dee05e869 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 540.132770] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Acquiring lock "a18427d3-449b-4589-b5a8-932d67d1620a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.133744] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Lock "a18427d3-449b-4589-b5a8-932d67d1620a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.133744] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg a1633b2e6d9c474a8da18d0e857ce01b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 540.143687] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a1633b2e6d9c474a8da18d0e857ce01b [ 540.159771] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 7f9279e635e24107b11f6a471f7672d3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 540.194031] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7f9279e635e24107b11f6a471f7672d3 [ 540.359396] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.720s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 540.360176] env[62599]: ERROR nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] Traceback (most recent call last): [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self.driver.spawn(context, instance, image_meta, [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self._vmops.spawn(context, instance, image_meta, injected_files, [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] vm_ref = self.build_virtual_machine(instance, [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] vif_infos = vmwarevif.get_vif_info(self._session, [ 540.360176] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] for vif in network_info: [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return self._sync_wrapper(fn, *args, **kwargs) [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self.wait() [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self[:] = self._gt.wait() [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return self._exit_event.wait() [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] result = hub.switch() [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 540.360546] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return self.greenlet.switch() [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] result = function(*args, **kwargs) [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] return func(*args, **kwargs) [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] raise e [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] nwinfo = self.network_api.allocate_for_instance( [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] created_port_ids = self._update_ports_for_instance( [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] with excutils.save_and_reraise_exception(): [ 540.360883] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] self.force_reraise() [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] raise self.value [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] updated_port = self._update_port( [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] _ensure_no_port_binding_failure(port) [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] raise exception.PortBindingFailed(port_id=port['id']) [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] nova.exception.PortBindingFailed: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. [ 540.361268] env[62599]: ERROR nova.compute.manager [instance: 030814bf-2510-49a7-a19a-b2afac554287] [ 540.361554] env[62599]: DEBUG nova.compute.utils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. {{(pid=62599) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 540.362093] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.547s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.370696] env[62599]: INFO nova.compute.claims [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 540.373004] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg ab60270ed78c46ad8cd6b864e8bb8fa9 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 540.376089] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Build of instance 030814bf-2510-49a7-a19a-b2afac554287 was re-scheduled: Binding failed for port d5e9ef63-ba42-4db9-9fb1-2c37885b2085, please check neutron logs for more information. {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2484}} [ 540.376978] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Unplugging VIFs for instance {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3010}} [ 540.377227] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Acquiring lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 540.377410] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Acquired lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 540.377596] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 540.377982] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg deb4f7c8c6e7430e8d1575b9c6df6b53 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 540.384199] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg deb4f7c8c6e7430e8d1575b9c6df6b53 [ 540.414805] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ab60270ed78c46ad8cd6b864e8bb8fa9 [ 540.417001] env[62599]: DEBUG oslo_concurrency.lockutils [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.545701] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394393, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026512} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 540.545701] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Deleted the datastore file {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 540.545880] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Moving file from [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to [datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039. {{(pid=62599) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 540.546047] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-1d82d2f3-591b-4227-a157-cadd6bf5efdf {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.552722] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 540.552722] env[62599]: value = "task-394394" [ 540.552722] env[62599]: _type = "Task" [ 540.552722] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 540.561944] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394394, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 540.621592] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Successfully created port: 79dd18ce-89a0-41a8-af78-14c44d9c3f0a {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 540.636279] env[62599]: DEBUG nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 540.638273] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 08039e91688745ddb472159209fd059b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 540.662636] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 540.695282] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 08039e91688745ddb472159209fd059b [ 540.699106] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 540.699317] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 540.699466] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 540.699645] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 540.699787] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 540.699928] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 540.700140] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 540.700288] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 540.700445] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 540.700602] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 540.700765] env[62599]: DEBUG nova.virt.hardware [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 540.702381] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc0ffb1a-a875-4208-bf5a-32c069b88eb2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.711005] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3eaf152-b8c3-4c41-92ec-308c4ed9456b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.875541] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 763d12ba7f43476b8a2d7350a25ad39a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 540.896810] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 763d12ba7f43476b8a2d7350a25ad39a [ 540.935662] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.073824] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394394, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.024692} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.074117] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] File moved {{(pid=62599) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 541.074310] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Cleaning up location [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27 {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 541.074469] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Deleting the datastore file [datastore1] vmware_temp/5e459fd2-df6a-4e42-a44f-a832c6093f27 {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 541.074775] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aba44f5b-6b46-42b0-9348-b790e91864d6 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.081438] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 541.081438] env[62599]: value = "task-394395" [ 541.081438] env[62599]: _type = "Task" [ 541.081438] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.090194] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394395, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.140323] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.140884] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg a3b477c7981f4732b3151c27a3026f90 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 541.154026] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a3b477c7981f4732b3151c27a3026f90 [ 541.161190] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.495341] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16fa967f-b005-4894-b610-9fcaffa8e398 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.505876] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8856f1-f2ac-4cf0-9adc-128902a3be11 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.554055] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cd836f9-09f5-4e34-bd4c-f9e7551d3f79 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.566156] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdfde3ff-3b6b-4a5e-b97e-d6586ee31fcf {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.582082] env[62599]: DEBUG nova.compute.provider_tree [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 541.582827] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 293e0d2f16944e5991e361e4c943e8b1 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 541.595415] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394395, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.026987} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 541.596064] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Deleted the datastore file {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 541.597663] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45913793-e8f6-43af-bc36-d03523be8df7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 541.603593] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 541.603593] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]52994540-a034-afd1-f712-7335998d9908" [ 541.603593] env[62599]: _type = "Task" [ 541.603593] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 541.613279] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]52994540-a034-afd1-f712-7335998d9908, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 541.615386] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 293e0d2f16944e5991e361e4c943e8b1 [ 541.645361] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Releasing lock "refresh_cache-030814bf-2510-49a7-a19a-b2afac554287" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 541.645858] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3033}} [ 541.646162] env[62599]: DEBUG nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 541.646496] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 541.689149] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 541.689149] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg b2b3e8eb880a40248174e97e94e09152 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 541.698706] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b2b3e8eb880a40248174e97e94e09152 [ 542.091521] env[62599]: DEBUG nova.scheduler.client.report [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 542.093939] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 12f5bcc0bbcb4675a6998dea53a66bde in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 542.110004] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 12f5bcc0bbcb4675a6998dea53a66bde [ 542.116050] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]52994540-a034-afd1-f712-7335998d9908, 'name': SearchDatastore_Task, 'duration_secs': 0.01034} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 542.116407] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Releasing lock "[datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 542.117534] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk to [datastore1] 76004611-5b01-4b15-9411-c60d404bfd25/76004611-5b01-4b15-9411-c60d404bfd25.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 542.118140] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4445ba31-adf4-40ad-97ec-267135855445 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.134123] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 542.134123] env[62599]: value = "task-394396" [ 542.134123] env[62599]: _type = "Task" [ 542.134123] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 542.143776] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394396, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.190330] env[62599]: DEBUG nova.network.neutron [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.190872] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 5038e7df42a6405386260e0976030638 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 542.229720] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5038e7df42a6405386260e0976030638 [ 542.600768] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 542.601397] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 542.603536] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg b517c0eb8ed44899bc9a764a4bbf8f7e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 542.611701] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.361s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.611701] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg ec1a7fa7aeef44dba0c6e18abe489d9c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 542.649153] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394396, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 542.657722] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b517c0eb8ed44899bc9a764a4bbf8f7e [ 542.661276] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ec1a7fa7aeef44dba0c6e18abe489d9c [ 542.704038] env[62599]: INFO nova.compute.manager [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] [instance: 030814bf-2510-49a7-a19a-b2afac554287] Took 1.05 seconds to deallocate network for instance. [ 542.704038] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg d8a1a89905bf47c6b17560254c4f0466 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 542.759296] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d8a1a89905bf47c6b17560254c4f0466 [ 543.110728] env[62599]: DEBUG nova.compute.utils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 543.111407] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg d250a8d90ac64967b68fb6ae5474a272 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 543.119588] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 543.119588] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 543.127834] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d250a8d90ac64967b68fb6ae5474a272 [ 543.158148] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394396, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.77723} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.158508] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk to [datastore1] 76004611-5b01-4b15-9411-c60d404bfd25/76004611-5b01-4b15-9411-c60d404bfd25.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 543.158734] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Extending root virtual disk to 1048576 {{(pid=62599) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 543.158984] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bf532d7e-8110-41fb-9265-0ebb05d110ae {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.169101] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 543.169101] env[62599]: value = "task-394397" [ 543.169101] env[62599]: _type = "Task" [ 543.169101] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.180168] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394397, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.210160] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg d63de7d723b44b96b41a3079992d9cda in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 543.261028] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d63de7d723b44b96b41a3079992d9cda [ 543.266948] env[62599]: ERROR nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. [ 543.266948] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.266948] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 543.266948] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.266948] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.266948] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.266948] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.266948] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.266948] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.266948] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 543.266948] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.266948] env[62599]: ERROR nova.compute.manager raise self.value [ 543.266948] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.266948] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.266948] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.266948] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.267404] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.267404] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.267404] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. [ 543.267404] env[62599]: ERROR nova.compute.manager [ 543.268127] env[62599]: Traceback (most recent call last): [ 543.268321] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.268321] env[62599]: listener.cb(fileno) [ 543.268423] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.268423] env[62599]: result = function(*args, **kwargs) [ 543.268819] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.268819] env[62599]: return func(*args, **kwargs) [ 543.268935] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 543.268935] env[62599]: raise e [ 543.269017] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 543.269017] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 543.269103] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.269103] env[62599]: created_port_ids = self._update_ports_for_instance( [ 543.269181] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.269181] env[62599]: with excutils.save_and_reraise_exception(): [ 543.269268] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.269268] env[62599]: self.force_reraise() [ 543.269360] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.269360] env[62599]: raise self.value [ 543.269444] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.269444] env[62599]: updated_port = self._update_port( [ 543.269522] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.269522] env[62599]: _ensure_no_port_binding_failure(port) [ 543.269599] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.269599] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.269688] env[62599]: nova.exception.PortBindingFailed: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. [ 543.269809] env[62599]: Removing descriptor: 16 [ 543.270823] env[62599]: ERROR nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Traceback (most recent call last): [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] yield resources [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self.driver.spawn(context, instance, image_meta, [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] vm_ref = self.build_virtual_machine(instance, [ 543.270823] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] for vif in network_info: [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return self._sync_wrapper(fn, *args, **kwargs) [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self.wait() [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self[:] = self._gt.wait() [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return self._exit_event.wait() [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.271143] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] result = hub.switch() [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return self.greenlet.switch() [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] result = function(*args, **kwargs) [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return func(*args, **kwargs) [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] raise e [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] nwinfo = self.network_api.allocate_for_instance( [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] created_port_ids = self._update_ports_for_instance( [ 543.271461] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] with excutils.save_and_reraise_exception(): [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self.force_reraise() [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] raise self.value [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] updated_port = self._update_port( [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] _ensure_no_port_binding_failure(port) [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] raise exception.PortBindingFailed(port_id=port['id']) [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] nova.exception.PortBindingFailed: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. [ 543.271772] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] [ 543.272855] env[62599]: INFO nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Terminating instance [ 543.279577] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquiring lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.279926] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquired lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.280367] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 543.281082] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 7ddd7a19b06f42c88368a8444b00e0f0 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 543.302370] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7ddd7a19b06f42c88368a8444b00e0f0 [ 543.388519] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc1d2c0-965f-4fc1-995e-a425893bcb1b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.397649] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a74129d4-0af4-4614-9b36-a2261f7925d6 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.445990] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddab24ee-32ad-45b1-94f2-65e332c73c29 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.458061] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7616f44b-901d-426c-80f3-a365da8a30de {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.480718] env[62599]: DEBUG nova.compute.provider_tree [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.481654] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 29d444b465bd4b7f81a263b743712bb4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 543.490032] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 29d444b465bd4b7f81a263b743712bb4 [ 543.616428] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 543.618162] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg a20ef6f251144b42a955d8390d98b136 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 543.633548] env[62599]: DEBUG nova.policy [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae1e8d03d42c4b84a314febcaf4c6705', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd4b1b4248728453eb942ccc0cb752e7e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 543.671170] env[62599]: DEBUG nova.compute.manager [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Received event network-changed-98a0d6aa-fdaf-42ed-9049-fec66cbc3237 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 543.671424] env[62599]: DEBUG nova.compute.manager [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Refreshing instance network info cache due to event network-changed-98a0d6aa-fdaf-42ed-9049-fec66cbc3237. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 543.671655] env[62599]: DEBUG oslo_concurrency.lockutils [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] Acquiring lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.677854] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a20ef6f251144b42a955d8390d98b136 [ 543.684561] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394397, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.079237} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 543.685046] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Extended root virtual disk {{(pid=62599) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 543.688582] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1d65c8-be96-4920-9081-d72f0d6fc50e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.720916] env[62599]: DEBUG nova.virt.vmwareapi.volumeops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 76004611-5b01-4b15-9411-c60d404bfd25/76004611-5b01-4b15-9411-c60d404bfd25.vmdk or device None with type sparse {{(pid=62599) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 543.724263] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e5b56677-bf3e-4f96-a3cf-4a74a9bbda80 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.746201] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 543.746201] env[62599]: value = "task-394398" [ 543.746201] env[62599]: _type = "Task" [ 543.746201] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 543.755553] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394398, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 543.767666] env[62599]: INFO nova.scheduler.client.report [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Deleted allocations for instance 030814bf-2510-49a7-a19a-b2afac554287 [ 543.774869] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Expecting reply to msg 5de8047bdefd46eda9aca6b0acec863a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 543.796300] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5de8047bdefd46eda9aca6b0acec863a [ 543.900672] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 543.984826] env[62599]: DEBUG nova.scheduler.client.report [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 543.987325] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg a24de6ddaf2247f5b5a8fd9d1283a3b9 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.003165] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a24de6ddaf2247f5b5a8fd9d1283a3b9 [ 544.122829] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg b10cce39ac164053b8a2e336d45281b8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.169087] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b10cce39ac164053b8a2e336d45281b8 [ 544.257275] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394398, 'name': ReconfigVM_Task, 'duration_secs': 0.350915} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.258522] env[62599]: DEBUG nova.virt.vmwareapi.volumeops [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 76004611-5b01-4b15-9411-c60d404bfd25/76004611-5b01-4b15-9411-c60d404bfd25.vmdk or device None with type sparse {{(pid=62599) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 544.259658] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b6036999-8473-4ca0-84d5-ab4eea35095a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.268055] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 544.268055] env[62599]: value = "task-394399" [ 544.268055] env[62599]: _type = "Task" [ 544.268055] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.282426] env[62599]: DEBUG oslo_concurrency.lockutils [None req-23509cf9-a46f-4ba6-bed8-04c656820dc8 tempest-ServerDiagnosticsTest-474362381 tempest-ServerDiagnosticsTest-474362381-project-member] Lock "030814bf-2510-49a7-a19a-b2afac554287" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.289s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.282935] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394399, 'name': Rename_Task} progress is 5%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.310148] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Acquiring lock "e010768e-56bd-4f35-a619-2ea9f2f2436f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.310526] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Lock "e010768e-56bd-4f35-a619-2ea9f2f2436f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.311088] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg 7bdb726528a34a84976e354fb5790970 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.326315] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7bdb726528a34a84976e354fb5790970 [ 544.419274] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.421953] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg f079cc6dacf74030944a0b7572806544 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.429230] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f079cc6dacf74030944a0b7572806544 [ 544.493882] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.889s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.494540] env[62599]: ERROR nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Traceback (most recent call last): [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self.driver.spawn(context, instance, image_meta, [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] vm_ref = self.build_virtual_machine(instance, [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] vif_infos = vmwarevif.get_vif_info(self._session, [ 544.494540] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] for vif in network_info: [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return self._sync_wrapper(fn, *args, **kwargs) [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self.wait() [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self[:] = self._gt.wait() [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return self._exit_event.wait() [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] result = hub.switch() [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 544.494890] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return self.greenlet.switch() [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] result = function(*args, **kwargs) [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] return func(*args, **kwargs) [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] raise e [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] nwinfo = self.network_api.allocate_for_instance( [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] created_port_ids = self._update_ports_for_instance( [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] with excutils.save_and_reraise_exception(): [ 544.495224] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] self.force_reraise() [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] raise self.value [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] updated_port = self._update_port( [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] _ensure_no_port_binding_failure(port) [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] raise exception.PortBindingFailed(port_id=port['id']) [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] nova.exception.PortBindingFailed: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. [ 544.495530] env[62599]: ERROR nova.compute.manager [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] [ 544.495805] env[62599]: DEBUG nova.compute.utils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. {{(pid=62599) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 544.499862] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.540s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.505993] env[62599]: INFO nova.compute.claims [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.508037] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 5082b7740fea4ad3a2662545ad5624b4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.513975] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Build of instance f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a was re-scheduled: Binding failed for port 86ad4574-224f-403d-a71f-544dee05e869, please check neutron logs for more information. {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2484}} [ 544.515602] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Unplugging VIFs for instance {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3010}} [ 544.516848] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Acquiring lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.516998] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Acquired lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.517171] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 544.517588] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 4ad2e02157f84239836f6e4812896c85 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.526176] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4ad2e02157f84239836f6e4812896c85 [ 544.569912] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5082b7740fea4ad3a2662545ad5624b4 [ 544.634343] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 544.676242] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 544.676242] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 544.676242] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 544.676461] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 544.676461] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 544.676461] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 544.676461] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 544.678423] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 544.678423] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 544.678423] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 544.678423] env[62599]: DEBUG nova.virt.hardware [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 544.678930] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9adc14c3-5945-4c4c-bf5b-e809eebefffe {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.689604] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9363693-7806-4bcb-86c6-28744d90575c {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.779991] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394399, 'name': Rename_Task, 'duration_secs': 0.163203} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 544.779991] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Powering on the VM {{(pid=62599) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 544.779991] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-cc1d4711-726b-4bcd-8d9e-c19e44acdbaf {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.788842] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 544.788842] env[62599]: value = "task-394400" [ 544.788842] env[62599]: _type = "Task" [ 544.788842] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 544.798389] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394400, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 544.813156] env[62599]: DEBUG nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 544.815844] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg a2428f13aac946cfa16f221cbba1db59 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.879073] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a2428f13aac946cfa16f221cbba1db59 [ 544.928774] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Releasing lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.928774] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 544.928774] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 544.928774] env[62599]: DEBUG oslo_concurrency.lockutils [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] Acquired lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.928774] env[62599]: DEBUG nova.network.neutron [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Refreshing network info cache for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237 {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 544.929291] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] Expecting reply to msg 648a3332e026430fb563de5085074383 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 544.929291] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e800f623-bae0-4745-8af4-302e1bfac080 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.935779] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77bb4d7a-8e4c-4674-a056-4acc841eb081 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.946832] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 648a3332e026430fb563de5085074383 [ 544.960084] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 91d61cd2-a61f-4a8a-8868-3f1f5654e176 could not be found. [ 544.960318] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 544.960493] env[62599]: INFO nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Took 0.04 seconds to destroy the instance on the hypervisor. [ 544.960730] env[62599]: DEBUG oslo.service.loopingcall [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.961253] env[62599]: DEBUG nova.compute.manager [-] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 544.961351] env[62599]: DEBUG nova.network.neutron [-] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 545.022394] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg debce8e5f5da40e6a7b9011eba0fbdc2 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.034869] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg debce8e5f5da40e6a7b9011eba0fbdc2 [ 545.112322] env[62599]: DEBUG nova.network.neutron [-] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.113440] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 89f770b70afa40c1af18fd65339cf853 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.114354] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.126337] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 89f770b70afa40c1af18fd65339cf853 [ 545.298177] env[62599]: DEBUG oslo_vmware.api [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394400, 'name': PowerOnVM_Task, 'duration_secs': 0.448218} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 545.298452] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Powered on the VM {{(pid=62599) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 545.298652] env[62599]: INFO nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Took 8.64 seconds to spawn the instance on the hypervisor. [ 545.298898] env[62599]: DEBUG nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Checking state {{(pid=62599) _get_power_state /opt/stack/nova/nova/compute/manager.py:1793}} [ 545.299678] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a30292d3-6fc7-43fd-9ce9-9fd4cb13dea5 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.309494] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg fa11c1ad06654b78a15d4506b2c793d8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.341168] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.352343] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg fa11c1ad06654b78a15d4506b2c793d8 [ 545.497390] env[62599]: DEBUG nova.network.neutron [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 545.616848] env[62599]: DEBUG nova.network.neutron [-] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.617542] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg fc468f62f1074144b0f0b6820e272a18 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.624230] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.624230] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 84f583529bdd4261a8f0d102c651c788 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.628989] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg fc468f62f1074144b0f0b6820e272a18 [ 545.639212] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 84f583529bdd4261a8f0d102c651c788 [ 545.701878] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-249c47de-7ada-4295-99b8-36416019ecb0 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.711118] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e16b4a-69f1-4328-a8d7-3ef388a49ea0 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.745810] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7031d2a-d8b8-40e9-956f-16cd5927b50e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.753432] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb735c6e-ae8e-4b82-9202-379002e9e405 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.768699] env[62599]: DEBUG nova.compute.provider_tree [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 545.769224] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg ebec3e97692e4422b14fe7c8f3051f29 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.777795] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ebec3e97692e4422b14fe7c8f3051f29 [ 545.794959] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Successfully created port: 1f928dd7-4550-4eb5-a624-51f01598e574 {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 545.820156] env[62599]: INFO nova.compute.manager [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Took 13.98 seconds to build instance. [ 545.820276] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 885cfc6cd9944d33a6664ca3bf572d07 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.842144] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 885cfc6cd9944d33a6664ca3bf572d07 [ 545.885069] env[62599]: DEBUG nova.network.neutron [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.885606] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] Expecting reply to msg eb73a7b46dcf44538d1fbf3f482fbda3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 545.895322] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg eb73a7b46dcf44538d1fbf3f482fbda3 [ 546.120849] env[62599]: INFO nova.compute.manager [-] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Took 1.16 seconds to deallocate network for instance. [ 546.122581] env[62599]: DEBUG nova.compute.claims [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 546.122760] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.125339] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Releasing lock "refresh_cache-f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.125513] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3033}} [ 546.125668] env[62599]: DEBUG nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 546.125833] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 546.184363] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 546.184971] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg d6b75ce8ed8d4b32be10cfc974cc0288 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 546.200775] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d6b75ce8ed8d4b32be10cfc974cc0288 [ 546.273708] env[62599]: DEBUG nova.scheduler.client.report [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.276160] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 01f4af667962451da826a756e0764859 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 546.292166] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 01f4af667962451da826a756e0764859 [ 546.321944] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0109e1c4-ddda-4606-b280-c2908708b6ba tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Lock "76004611-5b01-4b15-9411-c60d404bfd25" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.487s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.388405] env[62599]: DEBUG oslo_concurrency.lockutils [req-ab413766-c3ee-4763-8739-143364895e0f req-278895ce-0a53-4f7a-985d-b3cfda84aee4 service nova] Releasing lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 546.564274] env[62599]: ERROR nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 546.564274] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 546.564274] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 546.564274] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 546.564274] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 546.564274] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 546.564274] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 546.564274] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 546.564274] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.564274] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 546.564274] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.564274] env[62599]: ERROR nova.compute.manager raise self.value [ 546.564274] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 546.564274] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 546.564274] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.564274] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 546.564875] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.564875] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 546.564875] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 546.564875] env[62599]: ERROR nova.compute.manager [ 546.564875] env[62599]: Traceback (most recent call last): [ 546.564875] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 546.564875] env[62599]: listener.cb(fileno) [ 546.564875] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.564875] env[62599]: result = function(*args, **kwargs) [ 546.564875] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.564875] env[62599]: return func(*args, **kwargs) [ 546.564875] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 546.564875] env[62599]: raise e [ 546.564875] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 546.564875] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 546.564875] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 546.564875] env[62599]: created_port_ids = self._update_ports_for_instance( [ 546.564875] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 546.564875] env[62599]: with excutils.save_and_reraise_exception(): [ 546.564875] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.564875] env[62599]: self.force_reraise() [ 546.564875] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.564875] env[62599]: raise self.value [ 546.564875] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 546.564875] env[62599]: updated_port = self._update_port( [ 546.564875] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.564875] env[62599]: _ensure_no_port_binding_failure(port) [ 546.564875] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.564875] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 546.565616] env[62599]: nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 546.565616] env[62599]: Removing descriptor: 15 [ 546.565616] env[62599]: ERROR nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] Traceback (most recent call last): [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] yield resources [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self.driver.spawn(context, instance, image_meta, [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 546.565616] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] vm_ref = self.build_virtual_machine(instance, [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] for vif in network_info: [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return self._sync_wrapper(fn, *args, **kwargs) [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self.wait() [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self[:] = self._gt.wait() [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return self._exit_event.wait() [ 546.565973] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] result = hub.switch() [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return self.greenlet.switch() [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] result = function(*args, **kwargs) [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return func(*args, **kwargs) [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] raise e [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] nwinfo = self.network_api.allocate_for_instance( [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 546.566409] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] created_port_ids = self._update_ports_for_instance( [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] with excutils.save_and_reraise_exception(): [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self.force_reraise() [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] raise self.value [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] updated_port = self._update_port( [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] _ensure_no_port_binding_failure(port) [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 546.566807] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] raise exception.PortBindingFailed(port_id=port['id']) [ 546.567131] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 546.567131] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] [ 546.567131] env[62599]: INFO nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Terminating instance [ 546.567568] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Acquiring lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 546.567853] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Acquired lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 546.568083] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 546.568581] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg dedb7cb3b7b8455ebb77cc9cc77522af in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 546.577443] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg dedb7cb3b7b8455ebb77cc9cc77522af [ 546.688771] env[62599]: DEBUG nova.network.neutron [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 546.689346] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 449136a36ac84d6bb5b672a701b37960 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 546.700192] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 449136a36ac84d6bb5b672a701b37960 [ 546.779425] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.283s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.779982] env[62599]: DEBUG nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 546.781715] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg a348a396009e4af99a95cd09948bf0ae in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 546.784147] env[62599]: DEBUG oslo_concurrency.lockutils [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 6.367s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.784644] env[62599]: DEBUG oslo_concurrency.lockutils [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.784858] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62599) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 546.785180] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.624s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.786782] env[62599]: INFO nova.compute.claims [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 546.788409] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 98926c5e006743f2a91316b31b10bc83 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 546.793283] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c06e49d-74a5-42b9-8181-d31820fb0d72 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.804133] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2413b156-04dc-400a-b230-15b0751d45ae {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.818225] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2647e94d-b3a2-47b5-9f8b-0e44e6565ff1 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.832807] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21d95bc3-23ca-43cc-8a8e-a927be8cb8db {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.840783] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a348a396009e4af99a95cd09948bf0ae [ 546.841382] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 98926c5e006743f2a91316b31b10bc83 [ 546.869050] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181736MB free_disk=129GB free_vcpus=48 pci_devices=None {{(pid=62599) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 546.869202] env[62599]: DEBUG oslo_concurrency.lockutils [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.123674] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 547.193697] env[62599]: INFO nova.compute.manager [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] [instance: f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a] Took 1.07 seconds to deallocate network for instance. [ 547.196166] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg f36c4e0a462342d9ad6737cc966f287c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 547.239383] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f36c4e0a462342d9ad6737cc966f287c [ 547.286648] env[62599]: DEBUG nova.compute.utils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.287600] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 3c51239a4c8346398bfaf6d516f7358b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 547.287944] env[62599]: DEBUG nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Not allocating networking since 'none' was specified. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1979}} [ 547.294195] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 04c9acc7db1946b8a7b7c365edaa6f6b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 547.301301] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3c51239a4c8346398bfaf6d516f7358b [ 547.308699] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 04c9acc7db1946b8a7b7c365edaa6f6b [ 547.345140] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 547.345140] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 0dad1c947b7e482e8a73d0316e196a0d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 547.354230] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0dad1c947b7e482e8a73d0316e196a0d [ 547.713773] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 588f8ccfdd8c472bb999705b3da6260f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 547.775740] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 588f8ccfdd8c472bb999705b3da6260f [ 547.788909] env[62599]: DEBUG nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 547.790645] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 206a0516085641ac8482d6edea5c83fc in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 547.839766] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 206a0516085641ac8482d6edea5c83fc [ 547.850565] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Releasing lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.851162] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 547.851455] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 547.852508] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b287e95e-fd5a-4fad-b567-6f6815f853f8 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.864283] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff18cfcb-ccea-4366-8a8f-f08d6a8a7755 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.907263] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a888fb44-7545-48be-9135-7eb690994cdf could not be found. [ 547.907494] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 547.907682] env[62599]: INFO nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Took 0.06 seconds to destroy the instance on the hypervisor. [ 547.908261] env[62599]: DEBUG oslo.service.loopingcall [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 547.911278] env[62599]: DEBUG nova.compute.manager [-] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 547.911278] env[62599]: DEBUG nova.network.neutron [-] [instance: a888fb44-7545-48be-9135-7eb690994cdf] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 548.007555] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24276032-02ed-4c20-b627-f0e37a3a99b4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.020472] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4317798f-7385-4d85-a759-60bea87e51cb {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.058162] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e07b622f-7742-4c48-b6f8-551740ded0de {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.066933] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28c2e269-1510-4003-9004-0185c1aac890 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.087175] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Acquiring lock "37b14c16-ff14-4957-b394-c61450cae174" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.087513] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Lock "37b14c16-ff14-4957-b394-c61450cae174" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.088612] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 2893ab082fdb4cfaa932bd7fa494afef in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.111039] env[62599]: DEBUG nova.compute.provider_tree [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 548.111298] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg b708bf1d7cb74ab58d1f038582953035 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.112473] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2893ab082fdb4cfaa932bd7fa494afef [ 548.112971] env[62599]: DEBUG nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 548.114659] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 170f95cb4e2f4a1290202e232c64eab0 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.119438] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Acquiring lock "159fd42c-397a-4e00-ae6c-923ad73b8634" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.119648] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Lock "159fd42c-397a-4e00-ae6c-923ad73b8634" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.120122] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg ffdb719a4cdd4ecbab03493fb96f9cfa in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.145394] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b708bf1d7cb74ab58d1f038582953035 [ 548.166310] env[62599]: DEBUG nova.compute.manager [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Received event network-changed-79dd18ce-89a0-41a8-af78-14c44d9c3f0a {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.166496] env[62599]: DEBUG nova.compute.manager [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Refreshing instance network info cache due to event network-changed-79dd18ce-89a0-41a8-af78-14c44d9c3f0a. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 548.166710] env[62599]: DEBUG oslo_concurrency.lockutils [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] Acquiring lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.166940] env[62599]: DEBUG oslo_concurrency.lockutils [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] Acquired lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.167006] env[62599]: DEBUG nova.network.neutron [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Refreshing network info cache for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 548.167407] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] Expecting reply to msg 0cce664d97bc4660870aa6424eb4040c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.172334] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ffdb719a4cdd4ecbab03493fb96f9cfa [ 548.175231] env[62599]: DEBUG nova.network.neutron [-] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.175464] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 0134a7fad16f45ad9a0aaf40a82abfe2 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.182426] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 170f95cb4e2f4a1290202e232c64eab0 [ 548.185792] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0cce664d97bc4660870aa6424eb4040c [ 548.196435] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0134a7fad16f45ad9a0aaf40a82abfe2 [ 548.223365] env[62599]: DEBUG nova.compute.manager [req-fd7f4280-bf20-4f94-902e-d74325e7f293 req-03cb1cd0-b8a1-47ba-819d-d8ae4a616532 service nova] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Received event network-vif-deleted-98a0d6aa-fdaf-42ed-9049-fec66cbc3237 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 548.250325] env[62599]: INFO nova.scheduler.client.report [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Deleted allocations for instance f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a [ 548.259326] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Expecting reply to msg 3ba904f11cfc4f0eb1ef68305d34e51f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.272745] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3ba904f11cfc4f0eb1ef68305d34e51f [ 548.296049] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 34f64f96f7a54949befe729d7cc0106a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.350478] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 34f64f96f7a54949befe729d7cc0106a [ 548.620378] env[62599]: DEBUG nova.scheduler.client.report [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.622627] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 8b7af40f29de4c2c9ba946b2b157a676 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.632904] env[62599]: DEBUG nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 548.634774] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg 0f81e94563e74aa4b59c7ea0873f0319 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.643763] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8b7af40f29de4c2c9ba946b2b157a676 [ 548.659782] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 548.677301] env[62599]: DEBUG nova.network.neutron [-] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.677850] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 2e5764bb9ed94b0db324c51d1f983291 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.686188] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0f81e94563e74aa4b59c7ea0873f0319 [ 548.686370] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2e5764bb9ed94b0db324c51d1f983291 [ 548.708923] env[62599]: DEBUG nova.network.neutron [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 548.761756] env[62599]: DEBUG oslo_concurrency.lockutils [None req-0debc9f9-81b6-442d-8411-3e0e594dea81 tempest-ServerDiagnosticsNegativeTest-1167150184 tempest-ServerDiagnosticsNegativeTest-1167150184-project-member] Lock "f3dfc3d7-eccf-4c25-bb74-4eba0aa3407a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.590s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.799410] env[62599]: DEBUG nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 548.829010] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 548.829303] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 548.829464] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 548.829648] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 548.829856] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 548.830030] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 548.830245] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 548.830402] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 548.830661] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 548.830741] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 548.830910] env[62599]: DEBUG nova.virt.hardware [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 548.831832] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25aef2c3-e202-42a5-a808-436ef58e144e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.840089] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2712625-5669-4b78-99c6-516244009a55 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.870315] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Instance VIF info [] {{(pid=62599) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 548.885210] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Creating folder: Project (acb475bd8f32414287803c824ff383af). Parent ref: group-v107225. {{(pid=62599) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 548.885210] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8c958667-37f3-4301-b659-9ccad242d195 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.894529] env[62599]: INFO nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Created folder: Project (acb475bd8f32414287803c824ff383af) in parent group-v107225. [ 548.894741] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Creating folder: Instances. Parent ref: group-v107229. {{(pid=62599) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 548.894986] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9fe46a07-dbc5-498a-be9e-2d40cad102a0 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.905391] env[62599]: INFO nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Created folder: Instances in parent group-v107229. [ 548.905630] env[62599]: DEBUG oslo.service.loopingcall [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 548.905822] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Creating VM on the ESX host {{(pid=62599) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 548.906017] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7261b79b-78ee-48f1-86c7-b9e29ddedc2f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.922468] env[62599]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 548.922468] env[62599]: value = "task-394403" [ 548.922468] env[62599]: _type = "Task" [ 548.922468] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 548.929800] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394403, 'name': CreateVM_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.931285] env[62599]: DEBUG nova.network.neutron [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 548.931754] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] Expecting reply to msg 9f50688a6e9e49f5a06bf3d49c4e486e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 548.947569] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9f50688a6e9e49f5a06bf3d49c4e486e [ 549.138630] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.353s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 549.139227] env[62599]: DEBUG nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 549.141155] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 284bead0025644208081d7a045728d8d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 549.145861] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.805s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.147429] env[62599]: INFO nova.compute.claims [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 549.149159] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg 0e1b89981da043f99cfa7c7dd6b5b5a7 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 549.176590] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.184941] env[62599]: INFO nova.compute.manager [-] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Took 1.27 seconds to deallocate network for instance. [ 549.188349] env[62599]: DEBUG nova.compute.claims [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 549.188644] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.191028] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 284bead0025644208081d7a045728d8d [ 549.200800] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0e1b89981da043f99cfa7c7dd6b5b5a7 [ 549.267220] env[62599]: ERROR nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. [ 549.267220] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.267220] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 549.267220] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.267220] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.267220] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.267220] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.267220] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.267220] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.267220] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 549.267220] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.267220] env[62599]: ERROR nova.compute.manager raise self.value [ 549.267220] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.267220] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.267220] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.267220] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.267651] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.267651] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.267651] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. [ 549.267651] env[62599]: ERROR nova.compute.manager [ 549.267651] env[62599]: Traceback (most recent call last): [ 549.267651] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.267651] env[62599]: listener.cb(fileno) [ 549.267651] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.267651] env[62599]: result = function(*args, **kwargs) [ 549.267651] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.267651] env[62599]: return func(*args, **kwargs) [ 549.267651] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 549.267651] env[62599]: raise e [ 549.267651] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 549.267651] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 549.267651] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.267651] env[62599]: created_port_ids = self._update_ports_for_instance( [ 549.267651] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.267651] env[62599]: with excutils.save_and_reraise_exception(): [ 549.267651] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.267651] env[62599]: self.force_reraise() [ 549.267651] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.267651] env[62599]: raise self.value [ 549.267651] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.267651] env[62599]: updated_port = self._update_port( [ 549.267651] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.267651] env[62599]: _ensure_no_port_binding_failure(port) [ 549.267651] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.267651] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.268433] env[62599]: nova.exception.PortBindingFailed: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. [ 549.268433] env[62599]: Removing descriptor: 18 [ 549.268433] env[62599]: ERROR nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Traceback (most recent call last): [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] yield resources [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self.driver.spawn(context, instance, image_meta, [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.268433] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] vm_ref = self.build_virtual_machine(instance, [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] for vif in network_info: [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return self._sync_wrapper(fn, *args, **kwargs) [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self.wait() [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self[:] = self._gt.wait() [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return self._exit_event.wait() [ 549.268756] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] result = hub.switch() [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return self.greenlet.switch() [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] result = function(*args, **kwargs) [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return func(*args, **kwargs) [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] raise e [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] nwinfo = self.network_api.allocate_for_instance( [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 549.269100] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] created_port_ids = self._update_ports_for_instance( [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] with excutils.save_and_reraise_exception(): [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self.force_reraise() [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] raise self.value [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] updated_port = self._update_port( [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] _ensure_no_port_binding_failure(port) [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.269707] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] raise exception.PortBindingFailed(port_id=port['id']) [ 549.270195] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] nova.exception.PortBindingFailed: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. [ 549.270195] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] [ 549.270195] env[62599]: INFO nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Terminating instance [ 549.270195] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Acquiring lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.270375] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Acquired lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.270409] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 549.270901] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 5216d62bbf0a4a0aa89a64f717513f49 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 549.280634] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5216d62bbf0a4a0aa89a64f717513f49 [ 549.433146] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394403, 'name': CreateVM_Task} progress is 99%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.438878] env[62599]: DEBUG oslo_concurrency.lockutils [req-a3d64288-7170-40d0-bdab-c05932edcfa6 req-15606928-5267-469c-aec5-87157603be4b service nova] Releasing lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.571842] env[62599]: DEBUG nova.compute.manager [None req-b589fd44-7cd9-47d6-aeee-cd6e9f925b97 tempest-ServerDiagnosticsV248Test-1292888604 tempest-ServerDiagnosticsV248Test-1292888604-project-admin] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Checking state {{(pid=62599) _get_power_state /opt/stack/nova/nova/compute/manager.py:1793}} [ 549.573116] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9fe4957-02c3-4500-9275-4f0d50d8d136 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.584469] env[62599]: INFO nova.compute.manager [None req-b589fd44-7cd9-47d6-aeee-cd6e9f925b97 tempest-ServerDiagnosticsV248Test-1292888604 tempest-ServerDiagnosticsV248Test-1292888604-project-admin] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Retrieving diagnostics [ 549.585542] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83774739-32b6-44be-9bc7-c432a2833ed5 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.647779] env[62599]: DEBUG nova.compute.utils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 549.648753] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 8a6bd106fadc4442b4f2875e86cd9d9b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 549.653516] env[62599]: DEBUG nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 549.653516] env[62599]: DEBUG nova.network.neutron [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 549.664515] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg 9dfd99be4e6a4526816157d4ef1ad93e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 549.666894] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8a6bd106fadc4442b4f2875e86cd9d9b [ 549.674045] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9dfd99be4e6a4526816157d4ef1ad93e [ 549.803556] env[62599]: DEBUG nova.policy [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e60c4a3ab4ca467f848864e1f585ec72', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '76f21b49d0a6418b91854d77c448235b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 549.817963] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 549.933928] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394403, 'name': CreateVM_Task} progress is 99%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.955138] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.955687] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 0f653e56fb39435699f650390cc8e564 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 549.965902] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0f653e56fb39435699f650390cc8e564 [ 550.153608] env[62599]: DEBUG nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 550.155563] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg c9b7c56611e8493fa7bf103579e1c7be in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 550.223109] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c9b7c56611e8493fa7bf103579e1c7be [ 550.363034] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc340ab9-56ba-4373-b757-d5c524c6c3b3 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.366801] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6a3f8ab-9dee-4a9c-b91c-7d94efe6ff91 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.404756] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64501f40-8b41-4b90-be04-96263c504a73 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.412128] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5da41d7-6f3a-4133-8f53-33f43d25d740 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.433085] env[62599]: DEBUG nova.compute.provider_tree [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 550.433653] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg a1644de5ef6d4e33a0a71af2f276a36b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 550.442299] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a1644de5ef6d4e33a0a71af2f276a36b [ 550.447440] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394403, 'name': CreateVM_Task, 'duration_secs': 1.336949} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.447604] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Created VM on the ESX host {{(pid=62599) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 550.448508] env[62599]: DEBUG oslo_vmware.service [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facc3ab0-f5c1-4bca-a99d-86c940478ab2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.455565] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.455565] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.455865] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 550.456430] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9aa67186-c26d-475d-a936-f2c48e9eb2f5 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.461988] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Releasing lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.462411] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 550.462598] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 550.462934] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 550.462934] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]5218644b-e660-ea64-3ff6-f8757716762e" [ 550.462934] env[62599]: _type = "Task" [ 550.462934] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.463102] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fc0301d6-e3f9-4324-9230-3b4dc03303a4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.475896] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]5218644b-e660-ea64-3ff6-f8757716762e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.480420] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a047d03-1a27-41cc-95db-61732f9a12ac {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.507243] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e could not be found. [ 550.507243] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 550.507243] env[62599]: INFO nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 550.507243] env[62599]: DEBUG oslo.service.loopingcall [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.507243] env[62599]: DEBUG nova.compute.manager [-] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 550.507243] env[62599]: DEBUG nova.network.neutron [-] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 550.547363] env[62599]: DEBUG nova.network.neutron [-] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 550.547476] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 174e4f71958645eab08e16e00bb23d87 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 550.560769] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 174e4f71958645eab08e16e00bb23d87 [ 550.669396] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg c636e8329bbe4054a66b6682ab925c1f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 550.710162] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c636e8329bbe4054a66b6682ab925c1f [ 550.750777] env[62599]: DEBUG oslo_concurrency.lockutils [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Acquiring lock "6c2e9faf-3379-4446-a9be-ad7c32ede732" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 550.751011] env[62599]: DEBUG oslo_concurrency.lockutils [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Lock "6c2e9faf-3379-4446-a9be-ad7c32ede732" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.751518] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg 0c76f9b84a894291ad135aebda0a7d69 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 550.763768] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0c76f9b84a894291ad135aebda0a7d69 [ 550.935567] env[62599]: DEBUG nova.network.neutron [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Successfully created port: 52776fd0-be47-4c5a-b089-78c4216d400a {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 550.939589] env[62599]: DEBUG nova.scheduler.client.report [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.942517] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg 3e5e7d254e7e4fb4bf2df84199a7810d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 550.960165] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3e5e7d254e7e4fb4bf2df84199a7810d [ 550.979746] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.979995] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Processing image 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 550.980233] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.980375] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.980543] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 550.980786] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-827f8b79-421d-43b9-b383-37188f33df15 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.004640] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 551.004914] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62599) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 551.005681] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a315017-a4cf-45a3-ad69-8f469e8f3a95 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.011876] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3b1f689b-aa30-4543-9d4a-a95b869b6819 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.017460] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 551.017460] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]52af4344-41fa-68ab-98ba-463e35c94b89" [ 551.017460] env[62599]: _type = "Task" [ 551.017460] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.025549] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]52af4344-41fa-68ab-98ba-463e35c94b89, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.051480] env[62599]: DEBUG nova.network.neutron [-] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.051977] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 5e96f73aa9824d399a56cb57dc83a7b6 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 551.060850] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5e96f73aa9824d399a56cb57dc83a7b6 [ 551.174617] env[62599]: DEBUG nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 551.204877] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 551.205115] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 551.205262] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 551.205435] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 551.205573] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 551.205712] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 551.205914] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 551.206060] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 551.206216] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 551.206384] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 551.206546] env[62599]: DEBUG nova.virt.hardware [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 551.207401] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-759e153b-2444-4a42-b853-049e905a5cdf {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.218208] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b546c560-9e6d-4915-a4cb-1cc03eb0de29 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.253254] env[62599]: DEBUG nova.compute.manager [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] [instance: 6c2e9faf-3379-4446-a9be-ad7c32ede732] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 551.255234] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg f5706f71527c4920b1948479ce084643 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 551.300507] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f5706f71527c4920b1948479ce084643 [ 551.448606] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.303s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 551.449125] env[62599]: DEBUG nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 551.451006] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg a96f7ad4562f419591e3fb11ef380e8c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 551.452062] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.329s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 551.453838] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 07be6a108eb74e54a6149e8ebf0a14c8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 551.498313] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 07be6a108eb74e54a6149e8ebf0a14c8 [ 551.498972] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a96f7ad4562f419591e3fb11ef380e8c [ 551.528535] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Preparing fetch location {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 551.528908] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Creating directory with path [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 551.529257] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3cb31c58-e75a-49e7-82b1-14f21e2d280e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.551584] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Created directory with path [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 551.551827] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Fetch image to [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 551.552027] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Downloading image file data 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk on the data store datastore2 {{(pid=62599) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 551.552904] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2606bf-775a-4668-83a6-ec8b0e1a6b4a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.555666] env[62599]: INFO nova.compute.manager [-] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Took 1.05 seconds to deallocate network for instance. [ 551.558429] env[62599]: DEBUG nova.compute.claims [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 551.558675] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.561935] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dd1e4de-b17a-4f03-8d26-75c5d51f331a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.574034] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2169e12-dc34-4f36-b470-76852bf82a0d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.612315] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932ad4f4-b4c5-45c8-85f2-bdb2e674db38 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.618802] env[62599]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c8a4cb89-b974-49d5-8e9c-97545364dc33 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.647713] env[62599]: DEBUG nova.virt.vmwareapi.images [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Downloading image file data 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to the data store datastore2 {{(pid=62599) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 551.705144] env[62599]: DEBUG oslo_vmware.rw_handles [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62599) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 551.799420] env[62599]: DEBUG oslo_concurrency.lockutils [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 551.960491] env[62599]: DEBUG nova.compute.utils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 551.960491] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg 5b731efda10745978124e8fb79444b01 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 551.974944] env[62599]: DEBUG nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 551.975072] env[62599]: DEBUG nova.network.neutron [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 551.977274] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5b731efda10745978124e8fb79444b01 [ 552.132039] env[62599]: DEBUG nova.policy [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'db635e39c3e34e7c88f4e20783988d42', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8412803d6f7140bcb4e0e8884530c427', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 552.142923] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78d3746b-d68c-4655-ad98-69b14c86cfa8 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.158862] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b1ca635-34fa-4607-8d42-c4e11663cc96 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.216717] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74e2371-9d0a-4a44-a835-f62d1cb0844e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.235957] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76e7ef38-c5cf-4ae8-ac8a-90e406704253 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.259594] env[62599]: DEBUG nova.compute.provider_tree [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.260347] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 7f6db48698c2407094e0a2545951c74c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 552.267975] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7f6db48698c2407094e0a2545951c74c [ 552.449396] env[62599]: DEBUG nova.compute.manager [req-eb797782-2bf7-4800-a74b-d1ff0b198ecf req-62dd3862-0052-4e6f-8fb1-5037016f21c5 service nova] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Received event network-vif-deleted-79dd18ce-89a0-41a8-af78-14c44d9c3f0a {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 552.460814] env[62599]: DEBUG oslo_vmware.rw_handles [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Completed reading data from the image iterator. {{(pid=62599) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 552.461029] env[62599]: DEBUG oslo_vmware.rw_handles [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62599) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 552.465151] env[62599]: DEBUG nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 552.467127] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg 8e167a50a54e4962ac39d4f0f7ad7efd in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 552.494585] env[62599]: DEBUG nova.compute.manager [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Received event network-changed-1f928dd7-4550-4eb5-a624-51f01598e574 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 552.494770] env[62599]: DEBUG nova.compute.manager [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Refreshing instance network info cache due to event network-changed-1f928dd7-4550-4eb5-a624-51f01598e574. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 552.494981] env[62599]: DEBUG oslo_concurrency.lockutils [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] Acquiring lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 552.495118] env[62599]: DEBUG oslo_concurrency.lockutils [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] Acquired lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 552.495278] env[62599]: DEBUG nova.network.neutron [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Refreshing network info cache for port 1f928dd7-4550-4eb5-a624-51f01598e574 {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 552.495686] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] Expecting reply to msg a937e2200ac347a990067cbde87f8847 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 552.504161] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a937e2200ac347a990067cbde87f8847 [ 552.518163] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8e167a50a54e4962ac39d4f0f7ad7efd [ 552.519332] env[62599]: DEBUG nova.virt.vmwareapi.images [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Downloaded image file data 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk on the data store datastore2 {{(pid=62599) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 552.520779] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Caching image {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 552.521009] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Copying Virtual Disk [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk to [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 552.522430] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-89959b06-0d57-4dd5-83ff-c7e4d9dd3548 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.532433] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 552.532433] env[62599]: value = "task-394404" [ 552.532433] env[62599]: _type = "Task" [ 552.532433] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.543690] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394404, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.767614] env[62599]: DEBUG nova.scheduler.client.report [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 552.770166] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg b781ae458ced4236813c11772fd25771 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 552.782630] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b781ae458ced4236813c11772fd25771 [ 552.972147] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg dfe3f90ed78c48918bfa2bc70efd07d2 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 553.018183] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg dfe3f90ed78c48918bfa2bc70efd07d2 [ 553.043513] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394404, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.054507] env[62599]: DEBUG nova.network.neutron [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.239404] env[62599]: DEBUG nova.network.neutron [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.239902] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] Expecting reply to msg a0bab3cc8e02434fa80853aa9451edb3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 553.248665] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a0bab3cc8e02434fa80853aa9451edb3 [ 553.276368] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.276368] env[62599]: ERROR nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. [ 553.276368] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Traceback (most recent call last): [ 553.276368] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 553.276368] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self.driver.spawn(context, instance, image_meta, [ 553.276368] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 553.276368] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self._vmops.spawn(context, instance, image_meta, injected_files, [ 553.276368] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 553.276368] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] vm_ref = self.build_virtual_machine(instance, [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] vif_infos = vmwarevif.get_vif_info(self._session, [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] for vif in network_info: [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return self._sync_wrapper(fn, *args, **kwargs) [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self.wait() [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self[:] = self._gt.wait() [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return self._exit_event.wait() [ 553.276954] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] result = hub.switch() [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return self.greenlet.switch() [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] result = function(*args, **kwargs) [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] return func(*args, **kwargs) [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] raise e [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] nwinfo = self.network_api.allocate_for_instance( [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 553.277362] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] created_port_ids = self._update_ports_for_instance( [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] with excutils.save_and_reraise_exception(): [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] self.force_reraise() [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] raise self.value [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] updated_port = self._update_port( [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] _ensure_no_port_binding_failure(port) [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 553.277761] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] raise exception.PortBindingFailed(port_id=port['id']) [ 553.278149] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] nova.exception.PortBindingFailed: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. [ 553.278149] env[62599]: ERROR nova.compute.manager [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] [ 553.278149] env[62599]: DEBUG nova.compute.utils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. {{(pid=62599) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 553.278149] env[62599]: DEBUG oslo_concurrency.lockutils [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 6.408s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.278352] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 9900f9d7c4e54f0daed59501aa835d7f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 553.279782] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Build of instance 91d61cd2-a61f-4a8a-8868-3f1f5654e176 was re-scheduled: Binding failed for port 98a0d6aa-fdaf-42ed-9049-fec66cbc3237, please check neutron logs for more information. {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2484}} [ 553.280216] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Unplugging VIFs for instance {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3010}} [ 553.280430] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquiring lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 553.280568] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquired lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 553.280719] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 553.281100] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 75f9854e04bc44178d0c1f2a124835fb in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 553.293066] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 75f9854e04bc44178d0c1f2a124835fb [ 553.322817] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9900f9d7c4e54f0daed59501aa835d7f [ 553.475816] env[62599]: DEBUG nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 553.514012] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 553.514266] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 553.514418] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 553.514607] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 553.514771] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 553.514921] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 553.515370] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 553.515370] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 553.515470] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 553.515626] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 553.515741] env[62599]: DEBUG nova.virt.hardware [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 553.516820] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d1e3590-28ad-4f1d-b316-abdcb9d436a7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.530000] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09443cf5-7b1f-47a7-a8ad-926fb365afaf {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.551278] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394404, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.770177} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 553.551472] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Copied Virtual Disk [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk to [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 553.551665] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Deleting the datastore file [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/tmp-sparse.vmdk {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 553.551925] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f8fd170c-4bc5-4566-82b2-c5a4fed3b017 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 553.557810] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 553.557810] env[62599]: value = "task-394405" [ 553.557810] env[62599]: _type = "Task" [ 553.557810] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 553.565539] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394405, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.741882] env[62599]: DEBUG oslo_concurrency.lockutils [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] Releasing lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 553.742154] env[62599]: DEBUG nova.compute.manager [req-01fc6bee-9f87-4de0-a936-46c186113612 req-8d3a35f3-0268-454a-bd99-a4c5bcefa914 service nova] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Received event network-vif-deleted-1f928dd7-4550-4eb5-a624-51f01598e574 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 553.785511] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 653405ea1c1a47d882467669f2f82e19 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 553.796236] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 653405ea1c1a47d882467669f2f82e19 [ 553.815868] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 553.946604] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 553.947134] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 4518079d63a641f5bff2c367050d123c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 553.961656] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4518079d63a641f5bff2c367050d123c [ 554.005290] env[62599]: ERROR nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 52776fd0-be47-4c5a-b089-78c4216d400a, please check neutron logs for more information. [ 554.005290] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 554.005290] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 554.005290] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 554.005290] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.005290] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 554.005290] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.005290] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 554.005290] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.005290] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 554.005290] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.005290] env[62599]: ERROR nova.compute.manager raise self.value [ 554.005290] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.005290] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 554.005290] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.005290] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 554.005774] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.005774] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 554.005774] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 52776fd0-be47-4c5a-b089-78c4216d400a, please check neutron logs for more information. [ 554.005774] env[62599]: ERROR nova.compute.manager [ 554.005774] env[62599]: Traceback (most recent call last): [ 554.005774] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 554.005774] env[62599]: listener.cb(fileno) [ 554.005774] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.005774] env[62599]: result = function(*args, **kwargs) [ 554.005774] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.005774] env[62599]: return func(*args, **kwargs) [ 554.005774] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 554.005774] env[62599]: raise e [ 554.005774] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 554.005774] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 554.005774] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.005774] env[62599]: created_port_ids = self._update_ports_for_instance( [ 554.005774] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.005774] env[62599]: with excutils.save_and_reraise_exception(): [ 554.005774] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.005774] env[62599]: self.force_reraise() [ 554.005774] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.005774] env[62599]: raise self.value [ 554.005774] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.005774] env[62599]: updated_port = self._update_port( [ 554.005774] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.005774] env[62599]: _ensure_no_port_binding_failure(port) [ 554.005774] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.005774] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 554.006529] env[62599]: nova.exception.PortBindingFailed: Binding failed for port 52776fd0-be47-4c5a-b089-78c4216d400a, please check neutron logs for more information. [ 554.006529] env[62599]: Removing descriptor: 18 [ 554.006529] env[62599]: ERROR nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 52776fd0-be47-4c5a-b089-78c4216d400a, please check neutron logs for more information. [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Traceback (most recent call last): [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] yield resources [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] self.driver.spawn(context, instance, image_meta, [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 554.006529] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] vm_ref = self.build_virtual_machine(instance, [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] vif_infos = vmwarevif.get_vif_info(self._session, [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] for vif in network_info: [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] return self._sync_wrapper(fn, *args, **kwargs) [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] self.wait() [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] self[:] = self._gt.wait() [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] return self._exit_event.wait() [ 554.006865] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] result = hub.switch() [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] return self.greenlet.switch() [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] result = function(*args, **kwargs) [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] return func(*args, **kwargs) [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] raise e [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] nwinfo = self.network_api.allocate_for_instance( [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 554.007201] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] created_port_ids = self._update_ports_for_instance( [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] with excutils.save_and_reraise_exception(): [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] self.force_reraise() [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] raise self.value [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] updated_port = self._update_port( [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] _ensure_no_port_binding_failure(port) [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 554.007554] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] raise exception.PortBindingFailed(port_id=port['id']) [ 554.007838] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] nova.exception.PortBindingFailed: Binding failed for port 52776fd0-be47-4c5a-b089-78c4216d400a, please check neutron logs for more information. [ 554.007838] env[62599]: ERROR nova.compute.manager [instance: a18427d3-449b-4589-b5a8-932d67d1620a] [ 554.007838] env[62599]: INFO nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Terminating instance [ 554.008668] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Acquiring lock "refresh_cache-a18427d3-449b-4589-b5a8-932d67d1620a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 554.008817] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Acquired lock "refresh_cache-a18427d3-449b-4589-b5a8-932d67d1620a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 554.008976] env[62599]: DEBUG nova.network.neutron [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 554.009387] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 35274c23ec3242f5853b5d6f0094de7f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 554.019257] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 35274c23ec3242f5853b5d6f0094de7f [ 554.067890] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394405, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024025} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.069517] env[62599]: DEBUG nova.network.neutron [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Successfully created port: f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 554.071448] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Deleted the datastore file {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 554.071657] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Moving file from [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 to [datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039. {{(pid=62599) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 554.071923] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-76483f14-d97f-4a62-bf7d-1c58e00ca521 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.081692] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 554.081692] env[62599]: value = "task-394406" [ 554.081692] env[62599]: _type = "Task" [ 554.081692] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.096380] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394406, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.309414] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance 76004611-5b01-4b15-9411-c60d404bfd25 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.310095] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 7080272f0a7049b8be6f53a02821e5f3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 554.336946] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7080272f0a7049b8be6f53a02821e5f3 [ 554.449723] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Releasing lock "refresh_cache-91d61cd2-a61f-4a8a-8868-3f1f5654e176" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 554.449973] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3033}} [ 554.450139] env[62599]: DEBUG nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 554.450302] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 554.470436] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.471039] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 706d7f49beb049899e07ca0988ae5abc in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 554.477778] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 706d7f49beb049899e07ca0988ae5abc [ 554.534270] env[62599]: DEBUG nova.network.neutron [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 554.592908] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394406, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.022764} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.592908] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] File moved {{(pid=62599) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 554.592908] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Cleaning up location [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989 {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 554.592908] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Deleting the datastore file [datastore2] vmware_temp/036abb1f-8b77-4d2e-94e1-49581fe3b989 {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 554.592908] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-87e278db-5228-40c5-8c63-ab3239936835 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.600300] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 554.600300] env[62599]: value = "task-394407" [ 554.600300] env[62599]: _type = "Task" [ 554.600300] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.606729] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394407, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.779730] env[62599]: DEBUG nova.network.neutron [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.780286] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg ce041915c1c14dd0bf323040297ec64b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 554.790275] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ce041915c1c14dd0bf323040297ec64b [ 554.817378] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance 91d61cd2-a61f-4a8a-8868-3f1f5654e176 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 554.817378] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance a888fb44-7545-48be-9135-7eb690994cdf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.817378] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.817378] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance 66c56e34-39ee-41a3-94dc-6fd26a2cd726 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.817594] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance a18427d3-449b-4589-b5a8-932d67d1620a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.817594] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance e010768e-56bd-4f35-a619-2ea9f2f2436f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 554.817594] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 0fa6025d5946402ea4d6ad2dcc6305d0 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 554.849733] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0fa6025d5946402ea4d6ad2dcc6305d0 [ 554.973470] env[62599]: DEBUG nova.network.neutron [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 554.974029] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 5ad0d4d0e9a94746bfc0f9173ba0f542 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 554.984625] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5ad0d4d0e9a94746bfc0f9173ba0f542 [ 555.110098] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394407, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.040985} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.110424] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Deleted the datastore file {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 555.111337] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b792d56f-5cfe-4e67-92e1-40196a0378e4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.118885] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 555.118885] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]528e64be-7dc0-36ab-b5c9-0a386e6e86a5" [ 555.118885] env[62599]: _type = "Task" [ 555.118885] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.128843] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]528e64be-7dc0-36ab-b5c9-0a386e6e86a5, 'name': SearchDatastore_Task, 'duration_secs': 0.008393} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.128843] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.129313] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk to [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 555.129313] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-690bf1ca-96a1-4f07-b043-de8d5bbefe1f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.137225] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 555.137225] env[62599]: value = "task-394408" [ 555.137225] env[62599]: _type = "Task" [ 555.137225] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.144966] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394408, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.283314] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Releasing lock "refresh_cache-a18427d3-449b-4589-b5a8-932d67d1620a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 555.283706] env[62599]: DEBUG nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 555.283918] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 555.284272] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-935af57e-e04a-4209-811e-ed65e79f842c {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.292653] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a704ee-93d3-4342-af6b-b0dd9519d8f6 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.320422] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance 37b14c16-ff14-4957-b394-c61450cae174 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.321197] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg f3e60e626ce541418091e292bf12bc35 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 555.322552] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a18427d3-449b-4589-b5a8-932d67d1620a could not be found. [ 555.322651] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 555.322794] env[62599]: INFO nova.compute.manager [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 555.323028] env[62599]: DEBUG oslo.service.loopingcall [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 555.323545] env[62599]: DEBUG nova.compute.manager [-] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 555.323588] env[62599]: DEBUG nova.network.neutron [-] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 555.333250] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f3e60e626ce541418091e292bf12bc35 [ 555.476822] env[62599]: INFO nova.compute.manager [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: 91d61cd2-a61f-4a8a-8868-3f1f5654e176] Took 1.03 seconds to deallocate network for instance. [ 555.478621] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 699400c2b96d43f2a9d56617cb4fb355 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 555.500024] env[62599]: DEBUG nova.network.neutron [-] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 555.500561] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg bf65054bd2114b48bcb3c2ada696fadd in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 555.516592] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bf65054bd2114b48bcb3c2ada696fadd [ 555.535775] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 699400c2b96d43f2a9d56617cb4fb355 [ 555.574731] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edd593e2-dd9a-4294-aa31-c10b5a110817 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849-project-member] Acquiring lock "cdeaa360-5e34-4dbf-83f4-f6ba557bf941" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.574933] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edd593e2-dd9a-4294-aa31-c10b5a110817 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849-project-member] Lock "cdeaa360-5e34-4dbf-83f4-f6ba557bf941" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.575519] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edd593e2-dd9a-4294-aa31-c10b5a110817 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849-project-member] Expecting reply to msg 6f54e027034d48e9b747fa4217152b73 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 555.590909] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6f54e027034d48e9b747fa4217152b73 [ 555.646349] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394408, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.476856} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.646604] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk to [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 555.646806] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Extending root virtual disk to 1048576 {{(pid=62599) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 555.647046] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3513acdc-a123-4b95-a7b4-c2669624ada2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.654425] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 555.654425] env[62599]: value = "task-394409" [ 555.654425] env[62599]: _type = "Task" [ 555.654425] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.664589] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394409, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.824736] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance 159fd42c-397a-4e00-ae6c-923ad73b8634 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 555.825348] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg c667055c1ef642c6820dda188aeceef1 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 555.838013] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c667055c1ef642c6820dda188aeceef1 [ 555.984839] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg aa8b43b63c044d1b84c4505b21598705 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 556.003494] env[62599]: DEBUG nova.network.neutron [-] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.004072] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg bf7a6bf13a0a495f86e4270ec6da4192 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 556.012587] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bf7a6bf13a0a495f86e4270ec6da4192 [ 556.032249] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg aa8b43b63c044d1b84c4505b21598705 [ 556.077986] env[62599]: DEBUG nova.compute.manager [None req-edd593e2-dd9a-4294-aa31-c10b5a110817 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849-project-member] [instance: cdeaa360-5e34-4dbf-83f4-f6ba557bf941] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 556.079870] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edd593e2-dd9a-4294-aa31-c10b5a110817 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849-project-member] Expecting reply to msg 851b01c4b14145718e9039f764f45914 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 556.126629] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 851b01c4b14145718e9039f764f45914 [ 556.164640] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394409, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.058633} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.164750] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Extended root virtual disk {{(pid=62599) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 556.165482] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e42d217-e5af-457a-bc0a-576493b0dbdc {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.187323] env[62599]: DEBUG nova.virt.vmwareapi.volumeops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk or device None with type sparse {{(pid=62599) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 556.187507] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-039834f5-531c-42ee-b6fe-2abb4f0128ac {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.206863] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 556.206863] env[62599]: value = "task-394410" [ 556.206863] env[62599]: _type = "Task" [ 556.206863] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.215244] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394410, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.327764] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Instance 6c2e9faf-3379-4446-a9be-ad7c32ede732 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62599) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 556.328111] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62599) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 556.328155] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62599) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 556.506156] env[62599]: INFO nova.compute.manager [-] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Took 1.18 seconds to deallocate network for instance. [ 556.509030] env[62599]: DEBUG nova.compute.claims [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 556.509640] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.528718] env[62599]: INFO nova.scheduler.client.report [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Deleted allocations for instance 91d61cd2-a61f-4a8a-8868-3f1f5654e176 [ 556.540925] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 166e34d45beb487d8c36cb749a4e439c in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 556.568101] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-514d671a-3a04-4d09-bc5d-3be701fd8894 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.571379] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 166e34d45beb487d8c36cb749a4e439c [ 556.577833] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21165e13-ed57-4b12-bc59-262c016756fd {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.624569] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc3a6c84-ea78-42af-81f1-914c2d2882e6 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.637129] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b25d626-39dd-4bfd-807e-72356b6e7bc7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.649627] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edd593e2-dd9a-4294-aa31-c10b5a110817 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849 tempest-FloatingIPsAssociationNegativeTestJSON-2044589849-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 556.676436] env[62599]: DEBUG nova.compute.provider_tree [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 556.676963] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 18dfebe380ad4ff284129492b90b4680 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 556.688794] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 18dfebe380ad4ff284129492b90b4680 [ 556.716215] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394410, 'name': ReconfigVM_Task, 'duration_secs': 0.293354} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.716369] env[62599]: DEBUG nova.virt.vmwareapi.volumeops [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk or device None with type sparse {{(pid=62599) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 556.717069] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1ffc59fb-857a-4aa7-8be2-31a6f12a72dd {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.722996] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 556.722996] env[62599]: value = "task-394411" [ 556.722996] env[62599]: _type = "Task" [ 556.722996] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 556.732086] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394411, 'name': Rename_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.827900] env[62599]: DEBUG nova.compute.manager [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Received event network-changed-52776fd0-be47-4c5a-b089-78c4216d400a {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 556.827900] env[62599]: DEBUG nova.compute.manager [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Refreshing instance network info cache due to event network-changed-52776fd0-be47-4c5a-b089-78c4216d400a. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 556.827900] env[62599]: DEBUG oslo_concurrency.lockutils [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] Acquiring lock "refresh_cache-a18427d3-449b-4589-b5a8-932d67d1620a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 556.827900] env[62599]: DEBUG oslo_concurrency.lockutils [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] Acquired lock "refresh_cache-a18427d3-449b-4589-b5a8-932d67d1620a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 556.827900] env[62599]: DEBUG nova.network.neutron [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Refreshing network info cache for port 52776fd0-be47-4c5a-b089-78c4216d400a {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 556.828330] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] Expecting reply to msg 3a63f51e8c1f4dbfa215ba9f05ecc1bf in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 556.833551] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3a63f51e8c1f4dbfa215ba9f05ecc1bf [ 557.051457] env[62599]: DEBUG oslo_concurrency.lockutils [None req-f8fe1e5f-bc6d-417e-8498-074566f8027b tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Lock "91d61cd2-a61f-4a8a-8868-3f1f5654e176" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.864s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.122016] env[62599]: DEBUG oslo_concurrency.lockutils [None req-7b414ce1-6655-4b54-be90-dc5838c2a163 tempest-ImagesTestJSON-251474915 tempest-ImagesTestJSON-251474915-project-member] Acquiring lock "5d7ec25a-d47d-4a78-bc4d-0218209784e0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 557.122249] env[62599]: DEBUG oslo_concurrency.lockutils [None req-7b414ce1-6655-4b54-be90-dc5838c2a163 tempest-ImagesTestJSON-251474915 tempest-ImagesTestJSON-251474915-project-member] Lock "5d7ec25a-d47d-4a78-bc4d-0218209784e0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 557.122747] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-7b414ce1-6655-4b54-be90-dc5838c2a163 tempest-ImagesTestJSON-251474915 tempest-ImagesTestJSON-251474915-project-member] Expecting reply to msg f5ae878cd4b04292a9cc8090bcc8fb68 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 557.140595] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f5ae878cd4b04292a9cc8090bcc8fb68 [ 557.208394] env[62599]: ERROR nova.scheduler.client.report [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] [req-8b21605e-1e15-428a-9114-2c791ddb29f5] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a361070e-e5e7-4d3d-83b7-c85d96b4b61c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-8b21605e-1e15-428a-9114-2c791ddb29f5"}]} [ 557.233245] env[62599]: DEBUG nova.scheduler.client.report [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Refreshing inventories for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 557.238551] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394411, 'name': Rename_Task} progress is 14%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.256504] env[62599]: DEBUG nova.scheduler.client.report [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Updating ProviderTree inventory for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 557.256697] env[62599]: DEBUG nova.compute.provider_tree [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 557.273096] env[62599]: DEBUG nova.scheduler.client.report [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Refreshing aggregate associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, aggregates: None {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 557.297152] env[62599]: DEBUG nova.scheduler.client.report [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Refreshing trait associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 557.483583] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8835ac8d-e462-43bd-9084-177e152697c1 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.504260] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3db5d5fd-37f8-485f-8cd6-e4c61e365b40 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.533858] env[62599]: DEBUG nova.network.neutron [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 557.536836] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1f78d0c-7eb1-44c1-8003-739be84a3cbf {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.546156] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc5ebe54-70b3-4308-88ff-a3399ecf7830 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.561062] env[62599]: DEBUG nova.compute.provider_tree [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 557.561606] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg 7207e8cda2824c078ac03253792a47ae in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 557.572609] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7207e8cda2824c078ac03253792a47ae [ 557.628431] env[62599]: DEBUG nova.compute.manager [None req-7b414ce1-6655-4b54-be90-dc5838c2a163 tempest-ImagesTestJSON-251474915 tempest-ImagesTestJSON-251474915-project-member] [instance: 5d7ec25a-d47d-4a78-bc4d-0218209784e0] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 557.630168] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-7b414ce1-6655-4b54-be90-dc5838c2a163 tempest-ImagesTestJSON-251474915 tempest-ImagesTestJSON-251474915-project-member] Expecting reply to msg 75bffbb777b1423d9a8a25f9e948b7d9 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 557.663363] env[62599]: ERROR nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b, please check neutron logs for more information. [ 557.663363] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 557.663363] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 557.663363] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 557.663363] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.663363] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 557.663363] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.663363] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 557.663363] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.663363] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 557.663363] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.663363] env[62599]: ERROR nova.compute.manager raise self.value [ 557.663363] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.663363] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 557.663363] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.663363] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 557.663910] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.663910] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 557.663910] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b, please check neutron logs for more information. [ 557.663910] env[62599]: ERROR nova.compute.manager [ 557.663910] env[62599]: Traceback (most recent call last): [ 557.663910] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 557.663910] env[62599]: listener.cb(fileno) [ 557.663910] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.663910] env[62599]: result = function(*args, **kwargs) [ 557.663910] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.663910] env[62599]: return func(*args, **kwargs) [ 557.663910] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 557.663910] env[62599]: raise e [ 557.663910] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 557.663910] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 557.663910] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.663910] env[62599]: created_port_ids = self._update_ports_for_instance( [ 557.663910] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.663910] env[62599]: with excutils.save_and_reraise_exception(): [ 557.663910] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.663910] env[62599]: self.force_reraise() [ 557.663910] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.663910] env[62599]: raise self.value [ 557.663910] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.663910] env[62599]: updated_port = self._update_port( [ 557.663910] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.663910] env[62599]: _ensure_no_port_binding_failure(port) [ 557.663910] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.663910] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 557.664875] env[62599]: nova.exception.PortBindingFailed: Binding failed for port f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b, please check neutron logs for more information. [ 557.664875] env[62599]: Removing descriptor: 15 [ 557.664875] env[62599]: ERROR nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b, please check neutron logs for more information. [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Traceback (most recent call last): [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] yield resources [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] self.driver.spawn(context, instance, image_meta, [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 557.664875] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] vm_ref = self.build_virtual_machine(instance, [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] vif_infos = vmwarevif.get_vif_info(self._session, [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] for vif in network_info: [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] return self._sync_wrapper(fn, *args, **kwargs) [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] self.wait() [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] self[:] = self._gt.wait() [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] return self._exit_event.wait() [ 557.665245] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] result = hub.switch() [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] return self.greenlet.switch() [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] result = function(*args, **kwargs) [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] return func(*args, **kwargs) [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] raise e [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] nwinfo = self.network_api.allocate_for_instance( [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 557.665736] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] created_port_ids = self._update_ports_for_instance( [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] with excutils.save_and_reraise_exception(): [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] self.force_reraise() [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] raise self.value [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] updated_port = self._update_port( [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] _ensure_no_port_binding_failure(port) [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 557.666182] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] raise exception.PortBindingFailed(port_id=port['id']) [ 557.666574] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] nova.exception.PortBindingFailed: Binding failed for port f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b, please check neutron logs for more information. [ 557.666574] env[62599]: ERROR nova.compute.manager [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] [ 557.666574] env[62599]: INFO nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Terminating instance [ 557.668680] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Acquiring lock "refresh_cache-e010768e-56bd-4f35-a619-2ea9f2f2436f" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 557.668858] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Acquired lock "refresh_cache-e010768e-56bd-4f35-a619-2ea9f2f2436f" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 557.669025] env[62599]: DEBUG nova.network.neutron [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 557.669459] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg c9ac248de99d4a02aba40bdbbf3f2f79 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 557.679152] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c9ac248de99d4a02aba40bdbbf3f2f79 [ 557.683448] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 75bffbb777b1423d9a8a25f9e948b7d9 [ 557.734165] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394411, 'name': Rename_Task, 'duration_secs': 0.858636} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 557.734430] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Powering on the VM {{(pid=62599) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 557.734656] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9bc5f5d5-8539-42c8-9388-0cac2f39f76a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 557.740770] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 557.740770] env[62599]: value = "task-394412" [ 557.740770] env[62599]: _type = "Task" [ 557.740770] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 557.748488] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394412, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 557.840943] env[62599]: DEBUG nova.network.neutron [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.840943] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] Expecting reply to msg aa09050f446f45a3a33fc48a2ded7d36 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 557.850834] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg aa09050f446f45a3a33fc48a2ded7d36 [ 558.104610] env[62599]: DEBUG nova.scheduler.client.report [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Updated inventory for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with generation 18 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 558.104610] env[62599]: DEBUG nova.compute.provider_tree [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Updating resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c generation from 18 to 19 during operation: update_inventory {{(pid=62599) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 558.104610] env[62599]: DEBUG nova.compute.provider_tree [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 558.106914] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Expecting reply to msg ef0e014853f24003bea723fa17770944 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 558.121994] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ef0e014853f24003bea723fa17770944 [ 558.155788] env[62599]: DEBUG oslo_concurrency.lockutils [None req-7b414ce1-6655-4b54-be90-dc5838c2a163 tempest-ImagesTestJSON-251474915 tempest-ImagesTestJSON-251474915-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 558.238993] env[62599]: DEBUG nova.network.neutron [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 558.253490] env[62599]: DEBUG oslo_vmware.api [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394412, 'name': PowerOnVM_Task, 'duration_secs': 0.439959} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 558.254306] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Powered on the VM {{(pid=62599) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 558.254680] env[62599]: INFO nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Took 9.46 seconds to spawn the instance on the hypervisor. [ 558.255237] env[62599]: DEBUG nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Checking state {{(pid=62599) _get_power_state /opt/stack/nova/nova/compute/manager.py:1793}} [ 558.260721] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5dc5052-36b5-4aec-9220-e0eea1e996c3 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.273068] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 0a654dac7a674c3fa3b8ae27586cbdca in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 558.326258] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0a654dac7a674c3fa3b8ae27586cbdca [ 558.342279] env[62599]: DEBUG oslo_concurrency.lockutils [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] Releasing lock "refresh_cache-a18427d3-449b-4589-b5a8-932d67d1620a" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 558.342555] env[62599]: DEBUG nova.compute.manager [req-14928adc-018d-409b-9d1e-029bab4a63fb req-a7adf559-662a-4566-892d-6d5df6437728 service nova] [instance: a18427d3-449b-4589-b5a8-932d67d1620a] Received event network-vif-deleted-52776fd0-be47-4c5a-b089-78c4216d400a {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 558.468926] env[62599]: DEBUG nova.compute.manager [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Received event network-changed-f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 558.469126] env[62599]: DEBUG nova.compute.manager [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Refreshing instance network info cache due to event network-changed-f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 558.469313] env[62599]: DEBUG oslo_concurrency.lockutils [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] Acquiring lock "refresh_cache-e010768e-56bd-4f35-a619-2ea9f2f2436f" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 558.612085] env[62599]: DEBUG nova.compute.resource_tracker [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62599) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 558.612085] env[62599]: DEBUG oslo_concurrency.lockutils [None req-c9743e5b-4ba0-4153-bf97-f1647b3a4fbd None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 5.331s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 558.612085] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.949s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 558.612085] env[62599]: INFO nova.compute.claims [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 558.612085] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 73dd2b272dc4445ea92985e6f856e7d4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 558.631742] env[62599]: DEBUG nova.network.neutron [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 558.631742] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Expecting reply to msg 57e24728bff6469c8a19d994accbf624 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 558.644740] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 57e24728bff6469c8a19d994accbf624 [ 558.671163] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 73dd2b272dc4445ea92985e6f856e7d4 [ 558.784335] env[62599]: INFO nova.compute.manager [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Took 18.85 seconds to build instance. [ 558.784898] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg bf3abfba34a941c59e691644abb85221 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 558.801866] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bf3abfba34a941c59e691644abb85221 [ 559.116932] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 4d44568c2a9a439887672d957a5588ca in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 559.126129] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4d44568c2a9a439887672d957a5588ca [ 559.135385] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Releasing lock "refresh_cache-e010768e-56bd-4f35-a619-2ea9f2f2436f" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 559.135818] env[62599]: DEBUG nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 559.136024] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 559.136396] env[62599]: DEBUG oslo_concurrency.lockutils [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] Acquired lock "refresh_cache-e010768e-56bd-4f35-a619-2ea9f2f2436f" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 559.136581] env[62599]: DEBUG nova.network.neutron [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Refreshing network info cache for port f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 559.137052] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] Expecting reply to msg 32d16d290fe1488688011dff5b36b983 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 559.137797] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5b93353-3982-4e5a-85e3-b9e0ec017823 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.146732] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b0b23e-8727-419c-a31c-1bf507e07e22 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.163250] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 32d16d290fe1488688011dff5b36b983 [ 559.177075] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e010768e-56bd-4f35-a619-2ea9f2f2436f could not be found. [ 559.177465] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 559.177755] env[62599]: INFO nova.compute.manager [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 559.178082] env[62599]: DEBUG oslo.service.loopingcall [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 559.178370] env[62599]: DEBUG nova.compute.manager [-] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 559.178486] env[62599]: DEBUG nova.network.neutron [-] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 559.221731] env[62599]: DEBUG nova.network.neutron [-] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.221731] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 5ed11301f2e74ce1b675c4591864b348 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 559.228523] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5ed11301f2e74ce1b675c4591864b348 [ 559.290471] env[62599]: DEBUG oslo_concurrency.lockutils [None req-eea673b7-1e35-41d9-abbe-fd74f8e8595e tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Lock "66c56e34-39ee-41a3-94dc-6fd26a2cd726" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.365s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.724069] env[62599]: DEBUG nova.network.neutron [-] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 559.724584] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 7e43e6de3ee84398a0d66a00ffb5ad35 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 559.735403] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7e43e6de3ee84398a0d66a00ffb5ad35 [ 559.817230] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83676594-098a-4606-b635-9e4a6f754d33 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.822652] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dad98a89-0705-41ff-b2ae-de0c5b51c414 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.854942] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b764d0aa-8f9d-4bc2-be78-9ca8e6fd0888 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.862748] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-557d9297-0b0b-4136-8430-52e9426a4585 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.867465] env[62599]: DEBUG nova.network.neutron [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 559.879498] env[62599]: DEBUG nova.compute.provider_tree [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 559.880206] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 8ae7af8d23964b84a0c94a7b3ff11d37 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 559.894372] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8ae7af8d23964b84a0c94a7b3ff11d37 [ 560.142073] env[62599]: DEBUG nova.network.neutron [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 560.142615] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] Expecting reply to msg 4df500b78e684cb382e495da855afc34 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 560.152065] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4df500b78e684cb382e495da855afc34 [ 560.228495] env[62599]: INFO nova.compute.manager [-] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Took 1.05 seconds to deallocate network for instance. [ 560.229547] env[62599]: DEBUG nova.compute.claims [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 560.229547] env[62599]: DEBUG oslo_concurrency.lockutils [None req-1e061a8e-d36e-44ce-a8fd-079fd034a88e tempest-ServersTestManualDisk-1584989386 tempest-ServersTestManualDisk-1584989386-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.389023] env[62599]: DEBUG nova.scheduler.client.report [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 560.391421] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg cf0a4910882a4a1380cf84995eb3bb5d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 560.402319] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg cf0a4910882a4a1380cf84995eb3bb5d [ 560.645121] env[62599]: DEBUG oslo_concurrency.lockutils [req-3b3ceea6-5068-4fbe-a62d-4de735a36e39 req-91a74788-ebc2-453a-ac0a-46d2ac707d88 service nova] Releasing lock "refresh_cache-e010768e-56bd-4f35-a619-2ea9f2f2436f" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 560.894565] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.286s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 560.895145] env[62599]: DEBUG nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 560.897061] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 0e777f7d1d23421da3d65be74654618b in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 560.898087] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.722s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 560.899556] env[62599]: INFO nova.compute.claims [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 560.901253] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg f3384ebeb6b8424e87679914240ec428 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 560.943190] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0e777f7d1d23421da3d65be74654618b [ 560.954507] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f3384ebeb6b8424e87679914240ec428 [ 561.414032] env[62599]: DEBUG nova.compute.utils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 561.414698] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg a0b0b89c26b54827ac47fff7a3607177 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 561.416796] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg d40e7ad363ef404faccccf9712ae8fe9 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 561.420651] env[62599]: DEBUG nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 561.420651] env[62599]: DEBUG nova.network.neutron [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 561.428423] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d40e7ad363ef404faccccf9712ae8fe9 [ 561.429841] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a0b0b89c26b54827ac47fff7a3607177 [ 561.610120] env[62599]: DEBUG nova.policy [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '813526314627473cbd52b0d110eeaeb8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ab866d6bb9d84ccca2640d46f54e1b3d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 561.918405] env[62599]: DEBUG nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 561.920143] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 87e3f37f6de04e5e9aaa7c003b37f3a0 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 561.958145] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 87e3f37f6de04e5e9aaa7c003b37f3a0 [ 562.105668] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg afdf6f1488764b5cb58afed0bf55ca97 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 562.120018] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg afdf6f1488764b5cb58afed0bf55ca97 [ 562.124651] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18684dd-e18d-446a-a06a-35e2133b01bc {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.134848] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe13d67a-48e9-41d1-822a-adf6f396aa38 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.170394] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab67b7a8-463e-4da6-b74f-a96da0320795 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.180116] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ec938c-9c52-421e-9b65-22cec417a471 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.195806] env[62599]: DEBUG nova.compute.provider_tree [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 562.196354] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg a30221e241b44409b80b6d04e2440a2d in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 562.204462] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a30221e241b44409b80b6d04e2440a2d [ 562.428361] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 2c501d74d533404c8f78e6fd0bd836b4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 562.460221] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2c501d74d533404c8f78e6fd0bd836b4 [ 562.612904] env[62599]: INFO nova.compute.manager [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Rebuilding instance [ 562.657416] env[62599]: DEBUG nova.compute.manager [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Checking state {{(pid=62599) _get_power_state /opt/stack/nova/nova/compute/manager.py:1793}} [ 562.658398] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff9bc523-5c7a-4ce9-ad38-c95d75f46c90 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.667788] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg b1ba174726ce4284b8bf8cb75e54e505 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 562.699470] env[62599]: DEBUG nova.scheduler.client.report [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 562.702616] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg fb2869f3d9144e3b919f221615ac4266 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 562.715878] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg fb2869f3d9144e3b919f221615ac4266 [ 562.739804] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b1ba174726ce4284b8bf8cb75e54e505 [ 562.933090] env[62599]: DEBUG nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 562.956270] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 562.956270] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 562.956270] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 562.956489] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 562.956489] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 562.958607] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 562.958607] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 562.958607] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 562.958607] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 562.958607] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 562.958824] env[62599]: DEBUG nova.virt.hardware [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 562.958895] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9dfb92a-cc80-4692-ad79-7563b2d5306e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.969308] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f2a2ac8-2431-433d-9e08-60833abbc1ad {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.170562] env[62599]: DEBUG nova.network.neutron [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Successfully created port: 822c773b-7cbb-4caa-bb67-ab9ec51ef306 {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 563.175197] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Powering off the VM {{(pid=62599) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 563.175505] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d9608123-5178-4a8e-b346-146b36144c47 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.188343] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 563.188343] env[62599]: value = "task-394413" [ 563.188343] env[62599]: _type = "Task" [ 563.188343] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.197323] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394413, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.205349] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.307s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 563.205888] env[62599]: DEBUG nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 563.207877] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg a472483c72844b069dda55381e9b54fc in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 563.217094] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.028s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.220215] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 8c594c245a7a426dad3eea96b7818e05 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 563.248678] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a472483c72844b069dda55381e9b54fc [ 563.304903] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8c594c245a7a426dad3eea96b7818e05 [ 563.707255] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394413, 'name': PowerOffVM_Task, 'duration_secs': 0.117071} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 563.707548] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Powered off the VM {{(pid=62599) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 563.707899] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 563.711013] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92324917-c8ef-4894-9e13-ad4768fe8511 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.716522] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Unregistering the VM {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 563.718188] env[62599]: DEBUG nova.compute.utils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 563.718592] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg ead630480deb485e8ded1a901260d246 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 563.720501] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6123f14e-0e20-4e91-a0dc-d6cacbd3dc8b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.724561] env[62599]: DEBUG nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 563.724561] env[62599]: DEBUG nova.network.neutron [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 563.737568] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg ead630480deb485e8ded1a901260d246 [ 563.753884] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Unregistered the VM {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 563.753990] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Deleting contents of the VM from datastore datastore2 {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 563.754248] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Deleting the datastore file [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726 {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 563.754449] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-13899ffb-20f6-4a9a-a6aa-69fee021b6c5 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.761937] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 563.761937] env[62599]: value = "task-394415" [ 563.761937] env[62599]: _type = "Task" [ 563.761937] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.770735] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394415, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.851787] env[62599]: DEBUG nova.policy [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '23c7ae59d1f24d818fd39bbebda68da8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd7a62d6fa59a46da8d006aa721af7e67', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 563.897424] env[62599]: DEBUG oslo_concurrency.lockutils [None req-abc4ec41-64f7-45d3-b9e5-f47243550d20 tempest-ServersTestFqdnHostnames-445794756 tempest-ServersTestFqdnHostnames-445794756-project-member] Acquiring lock "56758047-1d5d-4600-b478-aec311fbd373" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 563.897697] env[62599]: DEBUG oslo_concurrency.lockutils [None req-abc4ec41-64f7-45d3-b9e5-f47243550d20 tempest-ServersTestFqdnHostnames-445794756 tempest-ServersTestFqdnHostnames-445794756-project-member] Lock "56758047-1d5d-4600-b478-aec311fbd373" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 563.898174] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-abc4ec41-64f7-45d3-b9e5-f47243550d20 tempest-ServersTestFqdnHostnames-445794756 tempest-ServersTestFqdnHostnames-445794756-project-member] Expecting reply to msg 2d14c5ab18224bd3a6ef2b1e7e2e921e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 563.910198] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2d14c5ab18224bd3a6ef2b1e7e2e921e [ 563.986668] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-021feb5d-b90d-46ef-b17e-c0715970446a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.994931] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28f0124-73b9-4272-b74f-963bb7c018dc {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.025216] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e90e0086-b72d-41e9-8000-241c6914262b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.032907] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70cba749-9fff-41a2-9b97-dc3edc598201 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.054141] env[62599]: DEBUG nova.compute.provider_tree [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 564.054691] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg e3307a2e545b411e96e0550f7318eed1 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 564.068952] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e3307a2e545b411e96e0550f7318eed1 [ 564.224097] env[62599]: DEBUG nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Start building block device mappings for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2868}} [ 564.224097] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg 5d8f0aa79765461c9e5bdc322e9785a6 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 564.267612] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5d8f0aa79765461c9e5bdc322e9785a6 [ 564.272078] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394415, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.104345} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.272557] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Deleted the datastore file {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 564.272858] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Deleted contents of the VM from datastore datastore2 {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 564.273151] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 564.275078] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 0c4beb471403499481441481b4d3f526 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 564.327596] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0c4beb471403499481441481b4d3f526 [ 564.379703] env[62599]: DEBUG nova.compute.manager [None req-90fc58bc-11d0-4eb0-8617-e03126a0f231 tempest-ServerDiagnosticsV248Test-1292888604 tempest-ServerDiagnosticsV248Test-1292888604-project-admin] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Checking state {{(pid=62599) _get_power_state /opt/stack/nova/nova/compute/manager.py:1793}} [ 564.381145] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85da7947-bf9d-4208-8eee-b96d873eb3e8 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.387770] env[62599]: INFO nova.compute.manager [None req-90fc58bc-11d0-4eb0-8617-e03126a0f231 tempest-ServerDiagnosticsV248Test-1292888604 tempest-ServerDiagnosticsV248Test-1292888604-project-admin] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Retrieving diagnostics [ 564.388535] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee7ee5c3-bb30-45fc-85ea-2c1345f7e7da {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.416672] env[62599]: DEBUG nova.compute.manager [None req-abc4ec41-64f7-45d3-b9e5-f47243550d20 tempest-ServersTestFqdnHostnames-445794756 tempest-ServersTestFqdnHostnames-445794756-project-member] [instance: 56758047-1d5d-4600-b478-aec311fbd373] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 564.418678] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-abc4ec41-64f7-45d3-b9e5-f47243550d20 tempest-ServersTestFqdnHostnames-445794756 tempest-ServersTestFqdnHostnames-445794756-project-member] Expecting reply to msg 6b78c79c3cfd46029c1188111ffb4062 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 564.459890] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6b78c79c3cfd46029c1188111ffb4062 [ 564.576294] env[62599]: ERROR nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [req-638a7bdc-e5b4-4924-94d3-db1d32acc072] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a361070e-e5e7-4d3d-83b7-c85d96b4b61c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-638a7bdc-e5b4-4924-94d3-db1d32acc072"}]}: nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 564.606402] env[62599]: DEBUG nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Refreshing inventories for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 564.624564] env[62599]: DEBUG nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Updating ProviderTree inventory for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 564.624867] env[62599]: DEBUG nova.compute.provider_tree [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 564.639363] env[62599]: DEBUG nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Refreshing aggregate associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, aggregates: None {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 564.658622] env[62599]: DEBUG nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Refreshing trait associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 564.731781] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg 344a8df59d9c45a7af46969e7612347a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 564.776469] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 344a8df59d9c45a7af46969e7612347a [ 564.780188] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 3c57e146400847c2a2529620c29724de in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 564.823832] env[62599]: DEBUG nova.network.neutron [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Successfully created port: 9264c1e7-f444-44ab-b040-4cde2fc2310e {{(pid=62599) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 564.833366] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3c57e146400847c2a2529620c29724de [ 564.862075] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b026627d-752d-4917-b353-a39f97a1e8d4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.870090] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85c9ec82-8801-4f35-97a6-32ce6dac2175 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.909884] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef29955-d5aa-4b14-a19c-29dad30326b2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.917846] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcabea6c-f924-4b7c-b175-7a39922ac38a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.936066] env[62599]: DEBUG nova.compute.provider_tree [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 564.936683] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg b2091233e3f54f77afc113266527ee48 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 564.942017] env[62599]: DEBUG oslo_concurrency.lockutils [None req-abc4ec41-64f7-45d3-b9e5-f47243550d20 tempest-ServersTestFqdnHostnames-445794756 tempest-ServersTestFqdnHostnames-445794756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.952242] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b2091233e3f54f77afc113266527ee48 [ 564.979010] env[62599]: DEBUG nova.compute.manager [req-451c56de-c4d6-4a29-a653-6916d794edf8 req-445ed205-29f7-4002-b7b3-f4839634a330 service nova] [instance: e010768e-56bd-4f35-a619-2ea9f2f2436f] Received event network-vif-deleted-f2a5e5f4-3c62-4859-9ebe-d885d2ef7c8b {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 565.235490] env[62599]: DEBUG nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Start spawning the instance on the hypervisor. {{(pid=62599) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2642}} [ 565.262287] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:26:57Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1857849942',id=23,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-380743535',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.262520] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.262678] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.262880] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.263047] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.263220] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.263425] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.263703] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.263879] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.264226] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.264529] env[62599]: DEBUG nova.virt.hardware [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.266037] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03e71463-7a0c-47e7-9999-c4621e3b7c90 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.274578] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1153ff1-c82a-4192-ba36-1bf1058d0bc5 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.311716] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-08-16T17:25:08Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-08-16T17:24:50Z,direct_url=,disk_format='vmdk',id=9647f31e-e8e8-4e9b-bd9d-aee2e7d10039,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='74f66d67c09640b285cb222a509a28de',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-08-16T17:24:51Z,virtual_size=,visibility=), allow threads: False {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 565.311928] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Flavor limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 565.312090] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Image limits 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 565.312267] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Flavor pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 565.312406] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Image pref 0:0:0 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 565.312547] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62599) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 565.312757] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 565.312925] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 565.313085] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Got 1 possible topologies {{(pid=62599) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 565.313227] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 565.313388] env[62599]: DEBUG nova.virt.hardware [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62599) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 565.314266] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a789114e-41e0-4749-af09-c3323f9badf6 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.322383] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed51c585-32f8-4568-af02-fb540ea3ea7c {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.337712] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Instance VIF info [] {{(pid=62599) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 565.343961] env[62599]: DEBUG oslo.service.loopingcall [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 565.344123] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Creating VM on the ESX host {{(pid=62599) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 565.344340] env[62599]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6ff2f8b2-f356-45c8-9b06-29e08b2922d4 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.361926] env[62599]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 565.361926] env[62599]: value = "task-394416" [ 565.361926] env[62599]: _type = "Task" [ 565.361926] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.369455] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394416, 'name': CreateVM_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.473416] env[62599]: DEBUG nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Updated inventory for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with generation 20 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 565.473554] env[62599]: DEBUG nova.compute.provider_tree [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Updating resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c generation from 20 to 21 during operation: update_inventory {{(pid=62599) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 565.473744] env[62599]: DEBUG nova.compute.provider_tree [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 565.479106] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg df20e9b7de914b769528ea4b33c05d65 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 565.495357] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg df20e9b7de914b769528ea4b33c05d65 [ 565.697096] env[62599]: ERROR nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 822c773b-7cbb-4caa-bb67-ab9ec51ef306, please check neutron logs for more information. [ 565.697096] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 565.697096] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 565.697096] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 565.697096] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.697096] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 565.697096] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.697096] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 565.697096] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.697096] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 565.697096] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.697096] env[62599]: ERROR nova.compute.manager raise self.value [ 565.697096] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.697096] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 565.697096] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.697096] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 565.697624] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.697624] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 565.697624] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 822c773b-7cbb-4caa-bb67-ab9ec51ef306, please check neutron logs for more information. [ 565.697624] env[62599]: ERROR nova.compute.manager [ 565.697624] env[62599]: Traceback (most recent call last): [ 565.697624] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 565.697624] env[62599]: listener.cb(fileno) [ 565.697624] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.697624] env[62599]: result = function(*args, **kwargs) [ 565.697624] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.697624] env[62599]: return func(*args, **kwargs) [ 565.697624] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 565.697624] env[62599]: raise e [ 565.697624] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 565.697624] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 565.697624] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.697624] env[62599]: created_port_ids = self._update_ports_for_instance( [ 565.697624] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.697624] env[62599]: with excutils.save_and_reraise_exception(): [ 565.697624] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.697624] env[62599]: self.force_reraise() [ 565.697624] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.697624] env[62599]: raise self.value [ 565.697624] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.697624] env[62599]: updated_port = self._update_port( [ 565.697624] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.697624] env[62599]: _ensure_no_port_binding_failure(port) [ 565.697624] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.697624] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 565.698433] env[62599]: nova.exception.PortBindingFailed: Binding failed for port 822c773b-7cbb-4caa-bb67-ab9ec51ef306, please check neutron logs for more information. [ 565.698433] env[62599]: Removing descriptor: 18 [ 565.698433] env[62599]: ERROR nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 822c773b-7cbb-4caa-bb67-ab9ec51ef306, please check neutron logs for more information. [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] Traceback (most recent call last): [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] yield resources [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] self.driver.spawn(context, instance, image_meta, [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.698433] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] vm_ref = self.build_virtual_machine(instance, [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] for vif in network_info: [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] return self._sync_wrapper(fn, *args, **kwargs) [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] self.wait() [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] self[:] = self._gt.wait() [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] return self._exit_event.wait() [ 565.698796] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] result = hub.switch() [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] return self.greenlet.switch() [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] result = function(*args, **kwargs) [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] return func(*args, **kwargs) [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] raise e [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] nwinfo = self.network_api.allocate_for_instance( [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.699223] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] created_port_ids = self._update_ports_for_instance( [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] with excutils.save_and_reraise_exception(): [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] self.force_reraise() [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] raise self.value [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] updated_port = self._update_port( [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] _ensure_no_port_binding_failure(port) [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.699713] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] raise exception.PortBindingFailed(port_id=port['id']) [ 565.700082] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] nova.exception.PortBindingFailed: Binding failed for port 822c773b-7cbb-4caa-bb67-ab9ec51ef306, please check neutron logs for more information. [ 565.700082] env[62599]: ERROR nova.compute.manager [instance: 37b14c16-ff14-4957-b394-c61450cae174] [ 565.700082] env[62599]: INFO nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Terminating instance [ 565.700371] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Acquiring lock "refresh_cache-37b14c16-ff14-4957-b394-c61450cae174" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.700507] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Acquired lock "refresh_cache-37b14c16-ff14-4957-b394-c61450cae174" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.700667] env[62599]: DEBUG nova.network.neutron [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 565.701085] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 294556c7072b4ccb8b06eb881778050e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 565.708358] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 294556c7072b4ccb8b06eb881778050e [ 565.874250] env[62599]: DEBUG oslo_vmware.api [-] Task: {'id': task-394416, 'name': CreateVM_Task, 'duration_secs': 0.288461} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.874598] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Created VM on the ESX host {{(pid=62599) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 565.875293] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.877552] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.878008] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 565.878314] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4550a8cb-8b28-4c89-a79b-312bd47cf72f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.884878] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 565.884878] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]52215046-0733-9356-1daf-4fd02d1fa4a7" [ 565.884878] env[62599]: _type = "Task" [ 565.884878] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.896244] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]52215046-0733-9356-1daf-4fd02d1fa4a7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.982126] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.765s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 565.982921] env[62599]: ERROR nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] Traceback (most recent call last): [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self.driver.spawn(context, instance, image_meta, [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self._vmops.spawn(context, instance, image_meta, injected_files, [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] vm_ref = self.build_virtual_machine(instance, [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] vif_infos = vmwarevif.get_vif_info(self._session, [ 565.982921] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] for vif in network_info: [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return self._sync_wrapper(fn, *args, **kwargs) [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self.wait() [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self[:] = self._gt.wait() [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return self._exit_event.wait() [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] result = hub.switch() [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 565.983335] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return self.greenlet.switch() [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] result = function(*args, **kwargs) [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] return func(*args, **kwargs) [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] raise e [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] nwinfo = self.network_api.allocate_for_instance( [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] created_port_ids = self._update_ports_for_instance( [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] with excutils.save_and_reraise_exception(): [ 565.983716] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] self.force_reraise() [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] raise self.value [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] updated_port = self._update_port( [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] _ensure_no_port_binding_failure(port) [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] raise exception.PortBindingFailed(port_id=port['id']) [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] nova.exception.PortBindingFailed: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. [ 565.984172] env[62599]: ERROR nova.compute.manager [instance: a888fb44-7545-48be-9135-7eb690994cdf] [ 565.984549] env[62599]: DEBUG nova.compute.utils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. {{(pid=62599) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 565.984982] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.426s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 565.986785] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg b8a076a8a61b4807ab730c2704eb81a5 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 565.998349] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Build of instance a888fb44-7545-48be-9135-7eb690994cdf was re-scheduled: Binding failed for port 79dd18ce-89a0-41a8-af78-14c44d9c3f0a, please check neutron logs for more information. {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2484}} [ 565.998510] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Unplugging VIFs for instance {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3010}} [ 565.998820] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Acquiring lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 565.998904] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Acquired lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 565.999042] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 565.999487] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg c1d82cb4915544e796646eb2629b2193 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 566.006928] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c1d82cb4915544e796646eb2629b2193 [ 566.038172] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b8a076a8a61b4807ab730c2704eb81a5 [ 566.229152] env[62599]: DEBUG nova.network.neutron [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.395462] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]52215046-0733-9356-1daf-4fd02d1fa4a7, 'name': SearchDatastore_Task, 'duration_secs': 0.008444} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.395782] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.395990] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Processing image 9647f31e-e8e8-4e9b-bd9d-aee2e7d10039 {{(pid=62599) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 566.396247] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 566.396387] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 566.396556] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 566.396823] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fde5ce28-3a6e-447b-bdfe-158c9a7dc9d2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.404454] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62599) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 566.405270] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62599) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 566.405389] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dded4864-053d-449b-ad1d-9eabfcb877ec {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.414129] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 566.414129] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]521fe3b8-1a3f-ce96-f0ab-3053ce63ee4d" [ 566.414129] env[62599]: _type = "Task" [ 566.414129] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.421183] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]521fe3b8-1a3f-ce96-f0ab-3053ce63ee4d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.451128] env[62599]: DEBUG nova.network.neutron [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.451546] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Expecting reply to msg 4a663e29e16b4773bf729f23d78ffa90 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 566.460420] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4a663e29e16b4773bf729f23d78ffa90 [ 566.526520] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 566.596436] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 566.599566] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 747ca044856745bdb7de7cd16e6cd77a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 566.612120] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 747ca044856745bdb7de7cd16e6cd77a [ 566.698424] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bc95f39-62bc-46b4-9221-c10618664259 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.706247] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff0f26d-d00a-4497-abf0-4d76c703b9ca {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.752418] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a9c8e8a-e023-496b-b444-9857480249ea {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.760425] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aefbaeb1-d6e3-4678-b0ae-5ab8ea81267f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.773318] env[62599]: DEBUG nova.compute.provider_tree [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 566.773924] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg e007687cb34c4d3d9823b3f945eed446 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 566.781892] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e007687cb34c4d3d9823b3f945eed446 [ 566.925650] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]521fe3b8-1a3f-ce96-f0ab-3053ce63ee4d, 'name': SearchDatastore_Task, 'duration_secs': 0.008211} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.926727] env[62599]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a01e77a1-5570-4436-b265-dafb7d8e2c3d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.933573] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 566.933573] env[62599]: value = "session[5273a345-7100-1b03-1f4f-2923726466a3]525540db-a8a7-032c-774c-4af8c25648eb" [ 566.933573] env[62599]: _type = "Task" [ 566.933573] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.945194] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': session[5273a345-7100-1b03-1f4f-2923726466a3]525540db-a8a7-032c-774c-4af8c25648eb, 'name': SearchDatastore_Task, 'duration_secs': 0.008331} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 566.945459] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.945707] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk to [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 566.945953] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4375eef9-994c-4394-b7fe-f73cfb298d78 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.952781] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 566.952781] env[62599]: value = "task-394417" [ 566.952781] env[62599]: _type = "Task" [ 566.952781] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 566.964087] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Releasing lock "refresh_cache-37b14c16-ff14-4957-b394-c61450cae174" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 566.964087] env[62599]: DEBUG nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 566.964087] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 566.964087] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9c407b0c-89f8-4f03-b9a1-e42dd648e692 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 566.972226] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394417, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.974721] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadde178-3521-4e2e-abe1-8397977bf36b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.003749] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 37b14c16-ff14-4957-b394-c61450cae174 could not be found. [ 567.004132] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 567.004520] env[62599]: INFO nova.compute.manager [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Took 0.04 seconds to destroy the instance on the hypervisor. [ 567.004873] env[62599]: DEBUG oslo.service.loopingcall [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 567.005174] env[62599]: DEBUG nova.compute.manager [-] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 567.005234] env[62599]: DEBUG nova.network.neutron [-] [instance: 37b14c16-ff14-4957-b394-c61450cae174] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.041986] env[62599]: DEBUG nova.network.neutron [-] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.042601] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 5daf111e063d4738b8dfefec3991efaf in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.060034] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 5daf111e063d4738b8dfefec3991efaf [ 567.105907] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Releasing lock "refresh_cache-a888fb44-7545-48be-9135-7eb690994cdf" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 567.106372] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3033}} [ 567.106631] env[62599]: DEBUG nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 567.106676] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 567.140331] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 567.141024] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg fc50c27d91cd435da49f756bba66094e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.150084] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg fc50c27d91cd435da49f756bba66094e [ 567.276784] env[62599]: DEBUG nova.scheduler.client.report [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 567.279340] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 36506ef7f34747c2a08f9d3ed4116a32 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.294888] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 36506ef7f34747c2a08f9d3ed4116a32 [ 567.462583] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394417, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.474027} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.462961] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039/9647f31e-e8e8-4e9b-bd9d-aee2e7d10039.vmdk to [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk {{(pid=62599) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 567.463065] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Extending root virtual disk to 1048576 {{(pid=62599) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 567.463307] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-73c58a06-364a-4547-b40b-4dc5d3f59582 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.470608] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 567.470608] env[62599]: value = "task-394422" [ 567.470608] env[62599]: _type = "Task" [ 567.470608] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.479875] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394422, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.498184] env[62599]: ERROR nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9264c1e7-f444-44ab-b040-4cde2fc2310e, please check neutron logs for more information. [ 567.498184] env[62599]: ERROR nova.compute.manager Traceback (most recent call last): [ 567.498184] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 567.498184] env[62599]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 567.498184] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.498184] env[62599]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 567.498184] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.498184] env[62599]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 567.498184] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.498184] env[62599]: ERROR nova.compute.manager self.force_reraise() [ 567.498184] env[62599]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.498184] env[62599]: ERROR nova.compute.manager raise self.value [ 567.498184] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.498184] env[62599]: ERROR nova.compute.manager updated_port = self._update_port( [ 567.498184] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.498184] env[62599]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 567.498692] env[62599]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.498692] env[62599]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 567.498692] env[62599]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9264c1e7-f444-44ab-b040-4cde2fc2310e, please check neutron logs for more information. [ 567.498692] env[62599]: ERROR nova.compute.manager [ 567.498692] env[62599]: Traceback (most recent call last): [ 567.498692] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 567.498692] env[62599]: listener.cb(fileno) [ 567.498692] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.498692] env[62599]: result = function(*args, **kwargs) [ 567.498692] env[62599]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.498692] env[62599]: return func(*args, **kwargs) [ 567.498692] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 567.498692] env[62599]: raise e [ 567.498692] env[62599]: File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 567.498692] env[62599]: nwinfo = self.network_api.allocate_for_instance( [ 567.498692] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.498692] env[62599]: created_port_ids = self._update_ports_for_instance( [ 567.498692] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.498692] env[62599]: with excutils.save_and_reraise_exception(): [ 567.498692] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.498692] env[62599]: self.force_reraise() [ 567.498692] env[62599]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.498692] env[62599]: raise self.value [ 567.498692] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.498692] env[62599]: updated_port = self._update_port( [ 567.498692] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.498692] env[62599]: _ensure_no_port_binding_failure(port) [ 567.498692] env[62599]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.498692] env[62599]: raise exception.PortBindingFailed(port_id=port['id']) [ 567.499508] env[62599]: nova.exception.PortBindingFailed: Binding failed for port 9264c1e7-f444-44ab-b040-4cde2fc2310e, please check neutron logs for more information. [ 567.499508] env[62599]: Removing descriptor: 15 [ 567.499704] env[62599]: ERROR nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9264c1e7-f444-44ab-b040-4cde2fc2310e, please check neutron logs for more information. [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Traceback (most recent call last): [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/compute/manager.py", line 2898, in _build_resources [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] yield resources [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] self.driver.spawn(context, instance, image_meta, [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] vm_ref = self.build_virtual_machine(instance, [ 567.499704] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] for vif in network_info: [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] return self._sync_wrapper(fn, *args, **kwargs) [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] self.wait() [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] self[:] = self._gt.wait() [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] return self._exit_event.wait() [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.500071] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] result = hub.switch() [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] return self.greenlet.switch() [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] result = function(*args, **kwargs) [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] return func(*args, **kwargs) [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] raise e [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] nwinfo = self.network_api.allocate_for_instance( [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] created_port_ids = self._update_ports_for_instance( [ 567.500849] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] with excutils.save_and_reraise_exception(): [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] self.force_reraise() [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] raise self.value [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] updated_port = self._update_port( [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] _ensure_no_port_binding_failure(port) [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] raise exception.PortBindingFailed(port_id=port['id']) [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] nova.exception.PortBindingFailed: Binding failed for port 9264c1e7-f444-44ab-b040-4cde2fc2310e, please check neutron logs for more information. [ 567.501244] env[62599]: ERROR nova.compute.manager [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] [ 567.501714] env[62599]: INFO nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Terminating instance [ 567.502501] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Acquiring lock "refresh_cache-159fd42c-397a-4e00-ae6c-923ad73b8634" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.502797] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Acquired lock "refresh_cache-159fd42c-397a-4e00-ae6c-923ad73b8634" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.503140] env[62599]: DEBUG nova.network.neutron [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 567.503897] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg 34b0c21fa0fd45639772e974b13e4311 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.544623] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 34b0c21fa0fd45639772e974b13e4311 [ 567.545218] env[62599]: DEBUG nova.network.neutron [-] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.545666] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 043144ccb59f436290d5c067ffaca526 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.557670] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 043144ccb59f436290d5c067ffaca526 [ 567.646044] env[62599]: DEBUG nova.network.neutron [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 567.646044] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg e7731030c26d4327a1e0a46b6500a966 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.656603] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e7731030c26d4327a1e0a46b6500a966 [ 567.784330] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.798s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 567.784330] env[62599]: ERROR nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. [ 567.784330] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Traceback (most recent call last): [ 567.784330] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/compute/manager.py", line 2645, in _build_and_run_instance [ 567.784330] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self.driver.spawn(context, instance, image_meta, [ 567.784330] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 539, in spawn [ 567.784330] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 567.784330] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 567.784330] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] vm_ref = self.build_virtual_machine(instance, [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] vif_infos = vmwarevif.get_vif_info(self._session, [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] for vif in network_info: [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return self._sync_wrapper(fn, *args, **kwargs) [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self.wait() [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self[:] = self._gt.wait() [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return self._exit_event.wait() [ 567.784959] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] result = hub.switch() [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return self.greenlet.switch() [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] result = function(*args, **kwargs) [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] return func(*args, **kwargs) [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/compute/manager.py", line 2013, in _allocate_network_async [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] raise e [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/compute/manager.py", line 1991, in _allocate_network_async [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] nwinfo = self.network_api.allocate_for_instance( [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 1229, in allocate_for_instance [ 567.785300] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] created_port_ids = self._update_ports_for_instance( [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 1365, in _update_ports_for_instance [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] with excutils.save_and_reraise_exception(): [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] self.force_reraise() [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] raise self.value [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 1340, in _update_ports_for_instance [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] updated_port = self._update_port( [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] _ensure_no_port_binding_failure(port) [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 567.785625] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] raise exception.PortBindingFailed(port_id=port['id']) [ 567.785926] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] nova.exception.PortBindingFailed: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. [ 567.785926] env[62599]: ERROR nova.compute.manager [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] [ 567.785926] env[62599]: DEBUG nova.compute.utils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. {{(pid=62599) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 567.785926] env[62599]: DEBUG oslo_concurrency.lockutils [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.987s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 567.790234] env[62599]: INFO nova.compute.claims [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] [instance: 6c2e9faf-3379-4446-a9be-ad7c32ede732] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 567.790234] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg d8fa9f26c8824356a5c0582da72e54e6 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.790592] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Build of instance b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e was re-scheduled: Binding failed for port 1f928dd7-4550-4eb5-a624-51f01598e574, please check neutron logs for more information. {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2484}} [ 567.790890] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Unplugging VIFs for instance {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3010}} [ 567.791071] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Acquiring lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 567.791212] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Acquired lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 567.791369] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 567.791750] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg c352ac6ce897469ba5d389b70ee54071 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 567.800902] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg c352ac6ce897469ba5d389b70ee54071 [ 567.828057] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d8fa9f26c8824356a5c0582da72e54e6 [ 567.980457] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394422, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068245} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.980686] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Extended root virtual disk {{(pid=62599) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 567.981456] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b162c06-3de9-4c9f-a1d4-653496d257d7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.000985] env[62599]: DEBUG nova.virt.vmwareapi.volumeops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Reconfiguring VM instance instance-00000007 to attach disk [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk or device None with type sparse {{(pid=62599) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 568.001267] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c7cd54ce-37bf-4e0e-95d5-0aa307156e3e {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.036144] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 568.036144] env[62599]: value = "task-394423" [ 568.036144] env[62599]: _type = "Task" [ 568.036144] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.047640] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394423, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.048177] env[62599]: INFO nova.compute.manager [-] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Took 1.04 seconds to deallocate network for instance. [ 568.050612] env[62599]: DEBUG nova.compute.claims [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 568.050742] env[62599]: DEBUG oslo_concurrency.lockutils [None req-edf53865-781f-4a06-9b35-98041e4c43b1 tempest-InstanceActionsV221TestJSON-1538959522 tempest-InstanceActionsV221TestJSON-1538959522-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.072233] env[62599]: DEBUG nova.network.neutron [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.147630] env[62599]: INFO nova.compute.manager [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] [instance: a888fb44-7545-48be-9135-7eb690994cdf] Took 1.04 seconds to deallocate network for instance. [ 568.151011] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg bd66d1b09d374df881d5698a71678ea8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.169872] env[62599]: DEBUG nova.network.neutron [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.169872] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Expecting reply to msg bba84659d24745649f96ad8001af11a8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.188168] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bba84659d24745649f96ad8001af11a8 [ 568.225104] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bd66d1b09d374df881d5698a71678ea8 [ 568.298364] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg 05e9ec0530fc4a7588210c62f53696be in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.306111] env[62599]: DEBUG oslo_concurrency.lockutils [None req-7acb4697-c5dc-4f39-b805-997aa8e77308 tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquiring lock "bd2f666e-b089-433b-b11b-37a13c8a3400" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.306490] env[62599]: DEBUG oslo_concurrency.lockutils [None req-7acb4697-c5dc-4f39-b805-997aa8e77308 tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Lock "bd2f666e-b089-433b-b11b-37a13c8a3400" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.307824] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 05e9ec0530fc4a7588210c62f53696be [ 568.336218] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.387452] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 568.387975] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 87dcb5912ec94087b620c134c9eabf95 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.398615] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 87dcb5912ec94087b620c134c9eabf95 [ 568.525891] env[62599]: DEBUG oslo_concurrency.lockutils [None req-341d551b-5fc3-45dd-b28a-28567cb9d6df tempest-AttachInterfacesUnderV243Test-1340435435 tempest-AttachInterfacesUnderV243Test-1340435435-project-member] Acquiring lock "cac55ba3-b856-40e5-b7f3-f3260dd51e00" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.526167] env[62599]: DEBUG oslo_concurrency.lockutils [None req-341d551b-5fc3-45dd-b28a-28567cb9d6df tempest-AttachInterfacesUnderV243Test-1340435435 tempest-AttachInterfacesUnderV243Test-1340435435-project-member] Lock "cac55ba3-b856-40e5-b7f3-f3260dd51e00" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.546475] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394423, 'name': ReconfigVM_Task, 'duration_secs': 0.279392} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.546826] env[62599]: DEBUG nova.virt.vmwareapi.volumeops [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Reconfigured VM instance instance-00000007 to attach disk [datastore2] 66c56e34-39ee-41a3-94dc-6fd26a2cd726/66c56e34-39ee-41a3-94dc-6fd26a2cd726.vmdk or device None with type sparse {{(pid=62599) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 568.547457] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c68b91b5-1868-4fa9-884e-7d3bee5a836f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.556634] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 568.556634] env[62599]: value = "task-394424" [ 568.556634] env[62599]: _type = "Task" [ 568.556634] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.565974] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394424, 'name': Rename_Task} progress is 5%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.657596] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 0c51f1eb8952402692deb13afbd6b1ad in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.668789] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg 9845b382697a49f5a990d62200c9bf6a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.672430] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Releasing lock "refresh_cache-159fd42c-397a-4e00-ae6c-923ad73b8634" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.672540] env[62599]: DEBUG nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 568.672721] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 568.673257] env[62599]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8a1921a4-9e6a-40d7-8599-1fc82ecf570b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.683305] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c21224-878c-4c10-b798-3c04c46b2b7c {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.696811] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9845b382697a49f5a990d62200c9bf6a [ 568.716649] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0c51f1eb8952402692deb13afbd6b1ad [ 568.717491] env[62599]: WARNING nova.virt.vmwareapi.vmops [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 159fd42c-397a-4e00-ae6c-923ad73b8634 could not be found. [ 568.717691] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 568.717866] env[62599]: INFO nova.compute.manager [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Took 0.05 seconds to destroy the instance on the hypervisor. [ 568.718101] env[62599]: DEBUG oslo.service.loopingcall [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 568.718658] env[62599]: DEBUG nova.compute.manager [-] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 568.718658] env[62599]: DEBUG nova.network.neutron [-] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.739842] env[62599]: DEBUG nova.compute.manager [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Received event network-changed-822c773b-7cbb-4caa-bb67-ab9ec51ef306 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 568.740296] env[62599]: DEBUG nova.compute.manager [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Refreshing instance network info cache due to event network-changed-822c773b-7cbb-4caa-bb67-ab9ec51ef306. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 568.740528] env[62599]: DEBUG oslo_concurrency.lockutils [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] Acquiring lock "refresh_cache-37b14c16-ff14-4957-b394-c61450cae174" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.740844] env[62599]: DEBUG oslo_concurrency.lockutils [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] Acquired lock "refresh_cache-37b14c16-ff14-4957-b394-c61450cae174" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.741016] env[62599]: DEBUG nova.network.neutron [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Refreshing network info cache for port 822c773b-7cbb-4caa-bb67-ab9ec51ef306 {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 568.741440] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] Expecting reply to msg e5c03b2391e8453e92b75a01b825e37a in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.750507] env[62599]: DEBUG nova.network.neutron [-] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.751038] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 4e7b78f70ab04e4dbf209b1998668bbd in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.758320] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg e5c03b2391e8453e92b75a01b825e37a [ 568.765140] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 4e7b78f70ab04e4dbf209b1998668bbd [ 568.813611] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5f8acf35-b7c3-4530-ab64-187e03f55d6e tempest-FloatingIPsAssociationTestJSON-2023467105 tempest-FloatingIPsAssociationTestJSON-2023467105-project-member] Acquiring lock "ccf5a2e6-cca9-4242-b2de-b1c29bbaa861" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 568.813611] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5f8acf35-b7c3-4530-ab64-187e03f55d6e tempest-FloatingIPsAssociationTestJSON-2023467105 tempest-FloatingIPsAssociationTestJSON-2023467105-project-member] Lock "ccf5a2e6-cca9-4242-b2de-b1c29bbaa861" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.890220] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Releasing lock "refresh_cache-b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 568.890453] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62599) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3033}} [ 568.890612] env[62599]: DEBUG nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 568.890780] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 568.908692] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 568.909261] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg a0d50b4bcbb042fb85f6eb77fdfe3fae in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 568.916653] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg a0d50b4bcbb042fb85f6eb77fdfe3fae [ 569.074310] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394424, 'name': Rename_Task, 'duration_secs': 0.137731} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.083047] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Powering on the VM {{(pid=62599) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 569.083539] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-743f3028-f085-4eeb-983b-5d44a24e9030 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.099139] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Waiting for the task: (returnval){ [ 569.099139] env[62599]: value = "task-394425" [ 569.099139] env[62599]: _type = "Task" [ 569.099139] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 569.108718] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcbf682-8f14-4d57-bb7c-5e1f8e1bc2b7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.116723] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394425, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 569.119452] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c0e6d7-aa23-46d5-823f-79a2e74355f2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.158710] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6104178-d911-41fc-98a9-702aaedc925c {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.166665] env[62599]: DEBUG oslo_concurrency.lockutils [None req-466e5880-f84e-40ef-8df2-f1961688cce9 tempest-ServerMetadataNegativeTestJSON-2090149976 tempest-ServerMetadataNegativeTestJSON-2090149976-project-member] Acquiring lock "4b313ce9-4fce-4eaf-968b-38a29402f46d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.172269] env[62599]: DEBUG oslo_concurrency.lockutils [None req-466e5880-f84e-40ef-8df2-f1961688cce9 tempest-ServerMetadataNegativeTestJSON-2090149976 tempest-ServerMetadataNegativeTestJSON-2090149976-project-member] Lock "4b313ce9-4fce-4eaf-968b-38a29402f46d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.172269] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquiring lock "76004611-5b01-4b15-9411-c60d404bfd25" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.172269] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Lock "76004611-5b01-4b15-9411-c60d404bfd25" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.172269] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquiring lock "76004611-5b01-4b15-9411-c60d404bfd25-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.172796] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Lock "76004611-5b01-4b15-9411-c60d404bfd25-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.172796] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Lock "76004611-5b01-4b15-9411-c60d404bfd25-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.175057] env[62599]: INFO nova.compute.manager [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Terminating instance [ 569.176755] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquiring lock "refresh_cache-76004611-5b01-4b15-9411-c60d404bfd25" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.176755] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Acquired lock "refresh_cache-76004611-5b01-4b15-9411-c60d404bfd25" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.176755] env[62599]: DEBUG nova.network.neutron [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Building network info cache for instance {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 569.176971] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg bbc6d118ab1c4226ba13b48c3b5d76fb in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.179797] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea02150-6ef9-4b34-b850-39d72811b7f2 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.185023] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bbc6d118ab1c4226ba13b48c3b5d76fb [ 569.187062] env[62599]: INFO nova.scheduler.client.report [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Deleted allocations for instance a888fb44-7545-48be-9135-7eb690994cdf [ 569.192380] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Expecting reply to msg 8fbd61c239d443ccb3c2cbf9d2e1b986 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.202295] env[62599]: DEBUG nova.compute.provider_tree [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 569.202841] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg 42986d938f7d4521a585a9ef3a8961a3 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.219325] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 42986d938f7d4521a585a9ef3a8961a3 [ 569.225678] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 8fbd61c239d443ccb3c2cbf9d2e1b986 [ 569.253412] env[62599]: DEBUG nova.network.neutron [-] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.253995] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg df36730384d547b0ae7a03d80e753ad8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.262751] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg df36730384d547b0ae7a03d80e753ad8 [ 569.327330] env[62599]: DEBUG nova.network.neutron [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.411817] env[62599]: DEBUG nova.network.neutron [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.412429] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 72b944f978f3487fba6529db536672a4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.422154] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 72b944f978f3487fba6529db536672a4 [ 569.613242] env[62599]: DEBUG oslo_vmware.api [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Task: {'id': task-394425, 'name': PowerOnVM_Task, 'duration_secs': 0.439196} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.613566] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Powered on the VM {{(pid=62599) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 569.613723] env[62599]: DEBUG nova.compute.manager [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Checking state {{(pid=62599) _get_power_state /opt/stack/nova/nova/compute/manager.py:1793}} [ 569.614504] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cd5a316-ce43-408e-9fc1-fa03960425a8 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.623914] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Expecting reply to msg 37e74001be1442dbb21c1496ee489b57 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.667125] env[62599]: DEBUG nova.network.neutron [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.667656] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] Expecting reply to msg 9eebb9c53f7948549029c4f4264c9747 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.678307] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9eebb9c53f7948549029c4f4264c9747 [ 569.693956] env[62599]: DEBUG oslo_concurrency.lockutils [None req-5c42214f-fc7f-449c-b422-b2f46fa20ed3 tempest-MigrationsAdminTest-326633083 tempest-MigrationsAdminTest-326633083-project-member] Lock "a888fb44-7545-48be-9135-7eb690994cdf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.204s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 569.694655] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-7acb4697-c5dc-4f39-b805-997aa8e77308 tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 50c64cf31abd4012935efb6e738efa7e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.699815] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 37e74001be1442dbb21c1496ee489b57 [ 569.711979] env[62599]: DEBUG nova.network.neutron [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 569.721595] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 50c64cf31abd4012935efb6e738efa7e [ 569.722828] env[62599]: ERROR nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] [req-ee725bf4-227f-4f30-afd0-ef2805c2aad9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a361070e-e5e7-4d3d-83b7-c85d96b4b61c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ee725bf4-227f-4f30-afd0-ef2805c2aad9"}]} [ 569.740602] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Refreshing inventories for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 569.756222] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Updating ProviderTree inventory for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 569.756571] env[62599]: DEBUG nova.compute.provider_tree [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 569.759152] env[62599]: INFO nova.compute.manager [-] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Took 1.04 seconds to deallocate network for instance. [ 569.767196] env[62599]: DEBUG nova.compute.claims [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Aborting claim: {{(pid=62599) abort /opt/stack/nova/nova/compute/claims.py:84}} [ 569.767420] env[62599]: DEBUG oslo_concurrency.lockutils [None req-6ac0a8c2-71f3-455b-ac4b-a735528ad4be tempest-ServersWithSpecificFlavorTestJSON-72491958 tempest-ServersWithSpecificFlavorTestJSON-72491958-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.770179] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Refreshing aggregate associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, aggregates: None {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 569.790870] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Refreshing trait associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 569.836392] env[62599]: DEBUG nova.compute.manager [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Received event network-changed-9264c1e7-f444-44ab-b040-4cde2fc2310e {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 569.836392] env[62599]: DEBUG nova.compute.manager [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Refreshing instance network info cache due to event network-changed-9264c1e7-f444-44ab-b040-4cde2fc2310e. {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11141}} [ 569.836392] env[62599]: DEBUG oslo_concurrency.lockutils [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] Acquiring lock "refresh_cache-159fd42c-397a-4e00-ae6c-923ad73b8634" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 569.836392] env[62599]: DEBUG oslo_concurrency.lockutils [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] Acquired lock "refresh_cache-159fd42c-397a-4e00-ae6c-923ad73b8634" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 569.836392] env[62599]: DEBUG nova.network.neutron [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Refreshing network info cache for port 9264c1e7-f444-44ab-b040-4cde2fc2310e {{(pid=62599) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 569.836690] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] Expecting reply to msg 6adf738293354e0ca8494eb889ac8453 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.844261] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 6adf738293354e0ca8494eb889ac8453 [ 569.906562] env[62599]: DEBUG nova.network.neutron [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.907598] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg f8d325c0a27645f789d72683af481eb5 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.915791] env[62599]: INFO nova.compute.manager [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] [instance: b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e] Took 1.02 seconds to deallocate network for instance. [ 569.918152] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg f24865da8e2e4489a4ffa5f2e55393fa in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 569.920837] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f8d325c0a27645f789d72683af481eb5 [ 569.960561] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg f24865da8e2e4489a4ffa5f2e55393fa [ 570.057774] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e861d39-9538-4ef5-8340-8c60349e1f1b {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.068339] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6299589-309c-448a-b841-488baedc2115 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.115151] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86301cee-29e0-4208-a059-b8ee2ab9880d {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.132203] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e90fd6-05a6-4a48-86eb-7ddd75919d3f {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.138908] env[62599]: DEBUG oslo_concurrency.lockutils [None req-e957c56c-6eb1-4f57-9b7a-9fc5bd9d4946 tempest-ServersAdmin275Test-1788286389 tempest-ServersAdmin275Test-1788286389-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.151383] env[62599]: DEBUG nova.compute.provider_tree [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 570.152156] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg 134c13da318144a6a85f50310aa20b16 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 570.160506] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 134c13da318144a6a85f50310aa20b16 [ 570.170005] env[62599]: DEBUG oslo_concurrency.lockutils [req-843b6e53-95e8-46ac-b0c2-c6bf151d5c19 req-ac8daa00-b987-4f8b-bee5-e7e493a5799b service nova] Releasing lock "refresh_cache-37b14c16-ff14-4957-b394-c61450cae174" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.200157] env[62599]: DEBUG nova.compute.manager [None req-7acb4697-c5dc-4f39-b805-997aa8e77308 tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] [instance: bd2f666e-b089-433b-b11b-37a13c8a3400] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 570.200157] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-7acb4697-c5dc-4f39-b805-997aa8e77308 tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Expecting reply to msg 2323ebfa5b734a8baa688cb05f6168b9 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 570.242119] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 2323ebfa5b734a8baa688cb05f6168b9 [ 570.384117] env[62599]: DEBUG nova.network.neutron [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 570.411240] env[62599]: DEBUG oslo_concurrency.lockutils [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Releasing lock "refresh_cache-76004611-5b01-4b15-9411-c60d404bfd25" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 570.411804] env[62599]: DEBUG nova.compute.manager [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Start destroying the instance on the hypervisor. {{(pid=62599) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3154}} [ 570.412086] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Destroying instance {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 570.413187] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19bd8158-eb62-4f1e-ad01-725a19ae45f7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.424098] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 49fed28a3a684575b44c82f68758ed56 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 570.425119] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Powering off the VM {{(pid=62599) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 570.425621] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-53431fc5-827e-4ef2-bb1c-e3122fd23fce {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.432471] env[62599]: DEBUG oslo_vmware.api [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 570.432471] env[62599]: value = "task-394427" [ 570.432471] env[62599]: _type = "Task" [ 570.432471] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 570.441408] env[62599]: DEBUG oslo_vmware.api [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394427, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 570.462097] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 49fed28a3a684575b44c82f68758ed56 [ 570.673368] env[62599]: ERROR nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] [req-2a0ac7d9-0b53-4943-9851-7220d5baabbe] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 128, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID a361070e-e5e7-4d3d-83b7-c85d96b4b61c. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-2a0ac7d9-0b53-4943-9851-7220d5baabbe"}]} [ 570.694318] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Refreshing inventories for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 570.702947] env[62599]: DEBUG nova.network.neutron [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] [instance: 159fd42c-397a-4e00-ae6c-923ad73b8634] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.703638] env[62599]: INFO oslo_messaging._drivers.amqpdriver [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] Expecting reply to msg 0de847c13274450a96ae08d8435d2398 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 570.715457] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0de847c13274450a96ae08d8435d2398 [ 570.716897] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Updating ProviderTree inventory for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 570.717151] env[62599]: DEBUG nova.compute.provider_tree [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Updating inventory in ProviderTree for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 570.726766] env[62599]: DEBUG oslo_concurrency.lockutils [None req-7acb4697-c5dc-4f39-b805-997aa8e77308 tempest-DeleteServersAdminTestJSON-718590941 tempest-DeleteServersAdminTestJSON-718590941-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.737226] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Refreshing aggregate associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, aggregates: 29884151-2b83-4422-b5bf-d435cd9eb9b0 {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 570.767021] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Refreshing trait associations for resource provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c, traits: COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=62599) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 570.948915] env[62599]: DEBUG oslo_vmware.api [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394427, 'name': PowerOffVM_Task, 'duration_secs': 0.154094} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 570.949384] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Powered off the VM {{(pid=62599) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 570.949818] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Unregistering the VM {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 570.950096] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-779dac96-2f7f-4b67-a644-4794e22e9831 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.954734] env[62599]: INFO nova.scheduler.client.report [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Deleted allocations for instance b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e [ 570.960559] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Expecting reply to msg 7d7e7a0727b94e7abab2f70b619aa198 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 570.977604] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 7d7e7a0727b94e7abab2f70b619aa198 [ 570.983153] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Unregistered the VM {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 570.983357] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Deleting contents of the VM from datastore datastore1 {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 570.983531] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Deleting the datastore file [datastore1] 76004611-5b01-4b15-9411-c60d404bfd25 {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 570.983819] env[62599]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f4c9584e-758c-4395-919c-a44d4f5eaef7 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 570.990086] env[62599]: DEBUG oslo_vmware.api [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for the task: (returnval){ [ 570.990086] env[62599]: value = "task-394429" [ 570.990086] env[62599]: _type = "Task" [ 570.990086] env[62599]: } to complete. {{(pid=62599) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 571.002114] env[62599]: DEBUG oslo_vmware.api [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394429, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 571.072757] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5849b4f8-f40f-46b0-bdcd-eeb47794fa78 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.085341] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a64f8e-b63d-43c7-9da9-39b31c070753 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.124009] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e7d359b-5811-42bd-b767-61f4c7d3f104 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.132358] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-189af845-94d4-4884-9a54-b1368c5d142a {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.152910] env[62599]: DEBUG nova.compute.provider_tree [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Inventory has not changed in ProviderTree for provider: a361070e-e5e7-4d3d-83b7-c85d96b4b61c {{(pid=62599) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.153082] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg 0e20591eaba343cf8624d663a5faf656 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 571.169507] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 0e20591eaba343cf8624d663a5faf656 [ 571.209441] env[62599]: DEBUG oslo_concurrency.lockutils [req-dde86d34-fac3-4916-9fce-f656f036d508 req-6dfe14e7-896e-45f2-b17e-540e8dde5437 service nova] Releasing lock "refresh_cache-159fd42c-397a-4e00-ae6c-923ad73b8634" {{(pid=62599) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 571.466949] env[62599]: DEBUG oslo_concurrency.lockutils [None req-d5585ecc-6d09-4edb-b158-1312929420b3 tempest-ServerExternalEventsTest-1876636801 tempest-ServerExternalEventsTest-1876636801-project-member] Lock "b23b9e5a-b4cd-43e0-ab66-3bdf1d82883e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.691s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 571.467563] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-341d551b-5fc3-45dd-b28a-28567cb9d6df tempest-AttachInterfacesUnderV243Test-1340435435 tempest-AttachInterfacesUnderV243Test-1340435435-project-member] Expecting reply to msg 675442f8b59a406ebef5bad5e63c6eb8 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 571.486412] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 675442f8b59a406ebef5bad5e63c6eb8 [ 571.500822] env[62599]: DEBUG oslo_vmware.api [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Task: {'id': task-394429, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090587} completed successfully. {{(pid=62599) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 571.501065] env[62599]: DEBUG nova.virt.vmwareapi.ds_util [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Deleted the datastore file {{(pid=62599) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 571.501241] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Deleted contents of the VM from datastore datastore1 {{(pid=62599) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 571.501403] env[62599]: DEBUG nova.virt.vmwareapi.vmops [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Instance destroyed {{(pid=62599) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 571.501568] env[62599]: INFO nova.compute.manager [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Took 1.09 seconds to destroy the instance on the hypervisor. [ 571.501829] env[62599]: DEBUG oslo.service.loopingcall [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62599) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 571.502009] env[62599]: DEBUG nova.compute.manager [-] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Deallocating network for instance {{(pid=62599) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2293}} [ 571.502102] env[62599]: DEBUG nova.network.neutron [-] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] deallocate_for_instance() {{(pid=62599) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 571.531140] env[62599]: DEBUG nova.network.neutron [-] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Instance cache missing network info. {{(pid=62599) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 571.531681] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg b683711872324146a80386726ca38fd5 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 571.540923] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg b683711872324146a80386726ca38fd5 [ 571.655543] env[62599]: DEBUG nova.scheduler.client.report [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Inventory has not changed for provider a361070e-e5e7-4d3d-83b7-c85d96b4b61c based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 129, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62599) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.657974] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg 95b696e25b534813bcfbef39b17458fc in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 571.671156] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 95b696e25b534813bcfbef39b17458fc [ 571.841155] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6f6418eb-5ab7-4c3f-949b-fdd9640a1d3c tempest-ServersAdmin275Test-1989415545 tempest-ServersAdmin275Test-1989415545-project-admin] Expecting reply to msg 3b7e062f2b564d5c8630f336ff8a7c86 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 571.851296] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 3b7e062f2b564d5c8630f336ff8a7c86 [ 571.970322] env[62599]: DEBUG nova.compute.manager [None req-341d551b-5fc3-45dd-b28a-28567cb9d6df tempest-AttachInterfacesUnderV243Test-1340435435 tempest-AttachInterfacesUnderV243Test-1340435435-project-member] [instance: cac55ba3-b856-40e5-b7f3-f3260dd51e00] Starting instance... {{(pid=62599) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2436}} [ 571.972386] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-341d551b-5fc3-45dd-b28a-28567cb9d6df tempest-AttachInterfacesUnderV243Test-1340435435 tempest-AttachInterfacesUnderV243Test-1340435435-project-member] Expecting reply to msg bfe89ebfb8de4894a281f4a5ce382095 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 572.012298] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg bfe89ebfb8de4894a281f4a5ce382095 [ 572.034864] env[62599]: DEBUG nova.network.neutron [-] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Updating instance_info_cache with network_info: [] {{(pid=62599) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 572.034864] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Expecting reply to msg 9b56d90134eb4e6eb7b4279a91772f71 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 572.043200] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 9b56d90134eb4e6eb7b4279a91772f71 [ 572.164353] env[62599]: DEBUG oslo_concurrency.lockutils [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 4.374s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.164353] env[62599]: DEBUG nova.compute.manager [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] [instance: 6c2e9faf-3379-4446-a9be-ad7c32ede732] Start building networks asynchronously for instance. {{(pid=62599) _build_resources /opt/stack/nova/nova/compute/manager.py:2833}} [ 572.164353] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg 1af1d72ec55f4b9590d9f224117841cb in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 572.164353] env[62599]: DEBUG oslo_concurrency.lockutils [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.654s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.166469] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-42960452-8150-4dd1-9d9d-f78f0c28c493 tempest-InstanceActionsTestJSON-1719799330 tempest-InstanceActionsTestJSON-1719799330-project-member] Expecting reply to msg 53b9842c30714022b802ec0d0a7258a4 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 572.191937] env[62599]: DEBUG nova.compute.manager [req-26d58b3b-7439-4208-bebb-a4513a5099d0 req-8a8bfb2d-65b3-4307-b762-4ebdd374e2e6 service nova] [instance: 37b14c16-ff14-4957-b394-c61450cae174] Received event network-vif-deleted-822c773b-7cbb-4caa-bb67-ab9ec51ef306 {{(pid=62599) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 572.215589] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 1af1d72ec55f4b9590d9f224117841cb [ 572.236966] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg 53b9842c30714022b802ec0d0a7258a4 [ 572.345133] env[62599]: INFO nova.compute.manager [None req-6f6418eb-5ab7-4c3f-949b-fdd9640a1d3c tempest-ServersAdmin275Test-1989415545 tempest-ServersAdmin275Test-1989415545-project-admin] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Rebuilding instance [ 572.388018] env[62599]: DEBUG nova.compute.manager [None req-6f6418eb-5ab7-4c3f-949b-fdd9640a1d3c tempest-ServersAdmin275Test-1989415545 tempest-ServersAdmin275Test-1989415545-project-admin] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Checking state {{(pid=62599) _get_power_state /opt/stack/nova/nova/compute/manager.py:1793}} [ 572.388905] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a560f8b1-40d0-4f9d-b433-ea25c6fae376 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.397874] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-6f6418eb-5ab7-4c3f-949b-fdd9640a1d3c tempest-ServersAdmin275Test-1989415545 tempest-ServersAdmin275Test-1989415545-project-admin] Expecting reply to msg db7fe176f7eb4245bc6e3e540fcd9c7e in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 572.446967] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg db7fe176f7eb4245bc6e3e540fcd9c7e [ 572.499468] env[62599]: DEBUG oslo_concurrency.lockutils [None req-341d551b-5fc3-45dd-b28a-28567cb9d6df tempest-AttachInterfacesUnderV243Test-1340435435 tempest-AttachInterfacesUnderV243Test-1340435435-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.536785] env[62599]: INFO nova.compute.manager [-] [instance: 76004611-5b01-4b15-9411-c60d404bfd25] Took 1.03 seconds to deallocate network for instance. [ 572.541231] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-266c858c-a494-4fc8-9bea-75390fb0cbd7 tempest-ServerDiagnosticsV248Test-1990900514 tempest-ServerDiagnosticsV248Test-1990900514-project-member] Expecting reply to msg d1b5423cddc14c9bb296f39488960e1f in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 572.594037] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d1b5423cddc14c9bb296f39488960e1f [ 572.674235] env[62599]: DEBUG nova.compute.utils [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Using /dev/sd instead of None {{(pid=62599) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.674235] env[62599]: INFO oslo_messaging._drivers.amqpdriver [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Expecting reply to msg d16de951d533411c96c8e7ce37dc4582 in queue reply_9dcc922abc674063ad7107323bf8efa6 [ 572.674235] env[62599]: DEBUG nova.compute.manager [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] [instance: 6c2e9faf-3379-4446-a9be-ad7c32ede732] Allocating IP information in the background. {{(pid=62599) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1983}} [ 572.674235] env[62599]: DEBUG nova.network.neutron [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] [instance: 6c2e9faf-3379-4446-a9be-ad7c32ede732] allocate_for_instance() {{(pid=62599) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 572.683312] env[62599]: INFO oslo_messaging._drivers.amqpdriver [-] Received RPC response for msg d16de951d533411c96c8e7ce37dc4582 [ 572.789988] env[62599]: DEBUG nova.policy [None req-a989ec5d-d2bd-4a84-a536-7843f480550f tempest-VolumesAssistedSnapshotsTest-672939131 tempest-VolumesAssistedSnapshotsTest-672939131-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4bef374b790f46a6ae364a9324da4c0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6161d613a52e41009e08b363d62812a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62599) authorize /opt/stack/nova/nova/policy.py:203}} [ 572.798961] env[62599]: DEBUG oslo_concurrency.lockutils [None req-561d5ac2-2bb0-4981-b795-3d69559767ad tempest-ServerActionsTestJSON-555375147 tempest-ServerActionsTestJSON-555375147-project-member] Acquiring lock "37b9cc98-cf29-412a-9902-f8fe5ed9448e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.798961] env[62599]: DEBUG oslo_concurrency.lockutils [None req-561d5ac2-2bb0-4981-b795-3d69559767ad tempest-ServerActionsTestJSON-555375147 tempest-ServerActionsTestJSON-555375147-project-member] Lock "37b9cc98-cf29-412a-9902-f8fe5ed9448e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62599) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.900364] env[62599]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f6418eb-5ab7-4c3f-949b-fdd9640a1d3c tempest-ServersAdmin275Test-1989415545 tempest-ServersAdmin275Test-1989415545-project-admin] [instance: 66c56e34-39ee-41a3-94dc-6fd26a2cd726] Powering off the VM {{(pid=62599) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 572.901368] env[62599]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bc91fe8-30f4-4faf-b456-2b017e128027 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.905419] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b45941b-4867-40af-881d-0eedf51e6034 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.912843] env[62599]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177cad65-a068-4462-9088-3ae8634a8067 {{(pid=62599) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.918260] env[62599]: DEBUG oslo_vmware.api [None req-6f6418eb-5ab7-4c3f-949b-fdd9640a1d3