[ 613.425051] env[69171]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=69171) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 613.425459] env[69171]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=69171) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 613.425459] env[69171]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=69171) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 613.425792] env[69171]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 613.523501] env[69171]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=69171) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 613.535022] env[69171]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.011s {{(pid=69171) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 614.167929] env[69171]: INFO nova.virt.driver [None req-2fec84e1-98d9-43a9-94ce-9106584db8f3 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 614.243224] env[69171]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.243450] env[69171]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.243539] env[69171]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=69171) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 617.583927] env[69171]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-27c6c902-d347-406b-b642-910d83b14a77 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.599881] env[69171]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=69171) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 617.600018] env[69171]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-677210cf-3b7d-467c-9858-0927bb4e91a2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.632693] env[69171]: INFO oslo_vmware.api [-] Successfully established new session; session ID is d0a8d. [ 617.632852] env[69171]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.389s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.633431] env[69171]: INFO nova.virt.vmwareapi.driver [None req-2fec84e1-98d9-43a9-94ce-9106584db8f3 None None] VMware vCenter version: 7.0.3 [ 617.636804] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f294925-ff04-469b-83f6-af5e46b3d022 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.658275] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c05f9cf-607e-44ac-a460-56e4dd329884 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.665397] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ef85a7b-4a78-456f-a782-2241ee8679cf {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.672365] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4489a94-b31f-4c3d-b250-e2d5740fc413 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.686219] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722f2819-fde8-40a6-92fe-1e9b50b728ba {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.692577] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ed92654-4f4b-4f78-9db3-f2d19d3cdd48 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.723437] env[69171]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-d5cd1e7b-8f77-42bf-ba94-388f43b7589f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.729145] env[69171]: DEBUG nova.virt.vmwareapi.driver [None req-2fec84e1-98d9-43a9-94ce-9106584db8f3 None None] Extension org.openstack.compute already exists. {{(pid=69171) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:225}} [ 617.732031] env[69171]: INFO nova.compute.provider_config [None req-2fec84e1-98d9-43a9-94ce-9106584db8f3 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 618.237070] env[69171]: DEBUG nova.context [None req-2fec84e1-98d9-43a9-94ce-9106584db8f3 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),4ba5d0ea-1c0a-4ac1-9ac7-6e10ef18a4a1(cell1) {{(pid=69171) load_cells /opt/stack/nova/nova/context.py:464}} [ 618.238510] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.238749] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.239475] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.239920] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Acquiring lock "4ba5d0ea-1c0a-4ac1-9ac7-6e10ef18a4a1" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 618.240127] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Lock "4ba5d0ea-1c0a-4ac1-9ac7-6e10ef18a4a1" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 618.241188] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Lock "4ba5d0ea-1c0a-4ac1-9ac7-6e10ef18a4a1" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 618.261802] env[69171]: INFO dbcounter [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Registered counter for database nova_cell0 [ 618.269908] env[69171]: INFO dbcounter [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Registered counter for database nova_cell1 [ 618.273068] env[69171]: DEBUG oslo_db.sqlalchemy.engines [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=69171) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 618.273431] env[69171]: DEBUG oslo_db.sqlalchemy.engines [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=69171) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 618.278100] env[69171]: ERROR nova.db.main.api [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.278100] env[69171]: result = function(*args, **kwargs) [ 618.278100] env[69171]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.278100] env[69171]: return func(*args, **kwargs) [ 618.278100] env[69171]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 618.278100] env[69171]: result = fn(*args, **kwargs) [ 618.278100] env[69171]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 618.278100] env[69171]: return f(*args, **kwargs) [ 618.278100] env[69171]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 618.278100] env[69171]: return db.service_get_minimum_version(context, binaries) [ 618.278100] env[69171]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 618.278100] env[69171]: _check_db_access() [ 618.278100] env[69171]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 618.278100] env[69171]: stacktrace = ''.join(traceback.format_stack()) [ 618.278100] env[69171]: [ 618.279138] env[69171]: ERROR nova.db.main.api [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 618.279138] env[69171]: result = function(*args, **kwargs) [ 618.279138] env[69171]: File "/opt/stack/nova/nova/utils.py", line 664, in context_wrapper [ 618.279138] env[69171]: return func(*args, **kwargs) [ 618.279138] env[69171]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 618.279138] env[69171]: result = fn(*args, **kwargs) [ 618.279138] env[69171]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 618.279138] env[69171]: return f(*args, **kwargs) [ 618.279138] env[69171]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 618.279138] env[69171]: return db.service_get_minimum_version(context, binaries) [ 618.279138] env[69171]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 618.279138] env[69171]: _check_db_access() [ 618.279138] env[69171]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 618.279138] env[69171]: stacktrace = ''.join(traceback.format_stack()) [ 618.279138] env[69171]: [ 618.279587] env[69171]: WARNING nova.objects.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 618.279756] env[69171]: WARNING nova.objects.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Failed to get minimum service version for cell 4ba5d0ea-1c0a-4ac1-9ac7-6e10ef18a4a1 [ 618.280163] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Acquiring lock "singleton_lock" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.280332] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Acquired lock "singleton_lock" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.280604] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Releasing lock "singleton_lock" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.280938] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Full set of CONF: {{(pid=69171) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 618.281096] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ******************************************************************************** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 618.281228] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Configuration options gathered from: {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 618.281370] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 618.281566] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 618.281696] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ================================================================================ {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 618.281911] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] allow_resize_to_same_host = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.282101] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] arq_binding_timeout = 300 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.282237] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] backdoor_port = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.282367] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] backdoor_socket = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.282537] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] block_device_allocate_retries = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.282703] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] block_device_allocate_retries_interval = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.282873] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cert = self.pem {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.283054] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.283230] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute_monitors = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.283399] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] config_dir = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.283572] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] config_drive_format = iso9660 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.283707] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.283870] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] config_source = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.284048] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] console_host = devstack {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.284219] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] control_exchange = nova {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.284379] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cpu_allocation_ratio = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.284541] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] daemon = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.284713] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] debug = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.284875] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] default_access_ip_network_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.285053] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] default_availability_zone = nova {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.285218] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] default_ephemeral_format = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.285383] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] default_green_pool_size = 1000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.285622] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.285788] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] default_schedule_zone = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.285950] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] disk_allocation_ratio = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.286130] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] enable_new_services = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.286315] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] enabled_apis = ['osapi_compute'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.286483] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] enabled_ssl_apis = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.286648] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] flat_injected = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.286815] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] force_config_drive = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.286977] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] force_raw_images = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.287166] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] graceful_shutdown_timeout = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.287334] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] heal_instance_info_cache_interval = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.287555] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] host = cpu-1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.287737] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] initial_cpu_allocation_ratio = 4.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.287905] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] initial_disk_allocation_ratio = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.288081] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] initial_ram_allocation_ratio = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.288307] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.288478] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instance_build_timeout = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.288642] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instance_delete_interval = 300 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.288810] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instance_format = [instance: %(uuid)s] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.288980] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instance_name_template = instance-%08x {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.289158] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instance_usage_audit = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.289355] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instance_usage_audit_period = month {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.289531] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.289699] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] instances_path = /opt/stack/data/nova/instances {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.289867] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] internal_service_availability_zone = internal {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.290099] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] key = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.290309] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] live_migration_retry_count = 30 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.290525] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_color = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.290702] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_config_append = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.290875] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.291051] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_dir = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.291219] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.291353] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_options = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.291520] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_rotate_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.291693] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_rotate_interval_type = days {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.291867] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] log_rotation_type = none {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.291995] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.292137] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.292309] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.292477] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.292608] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.292772] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] long_rpc_timeout = 1800 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.292935] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] max_concurrent_builds = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.293110] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] max_concurrent_live_migrations = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.293276] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] max_concurrent_snapshots = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.293440] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] max_local_block_devices = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.293602] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] max_logfile_count = 30 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.293764] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] max_logfile_size_mb = 200 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.293925] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] maximum_instance_delete_attempts = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.294114] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] metadata_listen = 0.0.0.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.294287] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] metadata_listen_port = 8775 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.294463] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] metadata_workers = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.294628] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] migrate_max_retries = -1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.294797] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] mkisofs_cmd = genisoimage {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.295013] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] my_block_storage_ip = 10.180.1.21 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.295160] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] my_ip = 10.180.1.21 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.295328] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] network_allocate_retries = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.295511] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.295682] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] osapi_compute_listen = 0.0.0.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.295849] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] osapi_compute_listen_port = 8774 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.296030] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] osapi_compute_unique_server_name_scope = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.296207] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] osapi_compute_workers = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.296380] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] password_length = 12 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.296541] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] periodic_enable = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.296705] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] periodic_fuzzy_delay = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.296876] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] pointer_model = usbtablet {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.297056] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] preallocate_images = none {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.297224] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] publish_errors = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.297359] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] pybasedir = /opt/stack/nova {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.297523] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ram_allocation_ratio = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.297684] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] rate_limit_burst = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.297855] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] rate_limit_except_level = CRITICAL {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.298034] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] rate_limit_interval = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.298206] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] reboot_timeout = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.298369] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] reclaim_instance_interval = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.298531] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] record = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.298701] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] reimage_timeout_per_gb = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.298870] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] report_interval = 120 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.299045] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] rescue_timeout = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.299212] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] reserved_host_cpus = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.299405] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] reserved_host_disk_mb = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.299573] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] reserved_host_memory_mb = 512 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.299737] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] reserved_huge_pages = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.299899] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] resize_confirm_window = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.300073] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] resize_fs_using_block_device = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.300239] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] resume_guests_state_on_host_boot = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.300452] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.300630] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] rpc_response_timeout = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.300797] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] run_external_periodic_tasks = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.300972] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] running_deleted_instance_action = reap {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.301152] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] running_deleted_instance_poll_interval = 1800 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.301318] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] running_deleted_instance_timeout = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.301479] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler_instance_sync_interval = 120 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.301660] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_down_time = 720 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.301826] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] servicegroup_driver = db {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.301985] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] shell_completion = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.302162] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] shelved_offload_time = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.302325] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] shelved_poll_interval = 3600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.302497] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] shutdown_timeout = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.302659] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] source_is_ipv6 = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.302820] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ssl_only = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.303079] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.303253] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] sync_power_state_interval = 600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.303420] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] sync_power_state_pool_size = 1000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.303592] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] syslog_log_facility = LOG_USER {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.303752] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] tempdir = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.303914] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] timeout_nbd = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.304102] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] transport_url = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.304271] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] update_resources_interval = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.304436] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] use_cow_images = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.304597] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] use_eventlog = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.304759] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] use_journal = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.304922] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] use_json = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.305094] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] use_rootwrap_daemon = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.305259] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] use_stderr = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.305422] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] use_syslog = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.305580] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vcpu_pin_set = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.305749] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plugging_is_fatal = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.305920] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plugging_timeout = 300 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.306102] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] virt_mkfs = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.306271] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] volume_usage_poll_interval = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.306436] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] watch_log_file = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.306630] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] web = /usr/share/spice-html5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 618.306840] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_concurrency.disable_process_locking = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.307154] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.307343] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.307515] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.307694] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_metrics.metrics_process_name = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.307866] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.308047] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.308241] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.auth_strategy = keystone {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.308412] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.compute_link_prefix = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.308591] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.308765] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.dhcp_domain = novalocal {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.308937] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.enable_instance_password = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.309116] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.glance_link_prefix = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.309326] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.309497] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.instance_list_cells_batch_strategy = distributed {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.309670] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.instance_list_per_project_cells = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.309836] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.list_records_by_skipping_down_cells = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.310013] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.local_metadata_per_cell = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.310190] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.max_limit = 1000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.310402] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.metadata_cache_expiration = 15 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.310609] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.neutron_default_tenant_id = default {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.310781] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.response_validation = warn {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.310953] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.use_neutron_default_nets = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.311142] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.311319] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.vendordata_dynamic_failure_fatal = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.311493] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.311676] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.vendordata_dynamic_ssl_certfile = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.311843] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.vendordata_dynamic_targets = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.312014] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.vendordata_jsonfile_path = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.312295] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api.vendordata_providers = ['StaticJSON'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.312398] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.backend = dogpile.cache.memcached {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.312570] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.backend_argument = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.312744] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.config_prefix = cache.oslo {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.312916] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.dead_timeout = 60.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.313096] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.debug_cache_backend = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.313266] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.enable_retry_client = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.313432] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.enable_socket_keepalive = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.313604] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.enabled = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.313770] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.enforce_fips_mode = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.313934] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.expiration_time = 600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.314118] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.hashclient_retry_attempts = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.314293] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.hashclient_retry_delay = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.314462] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_dead_retry = 300 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.314624] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_password = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.314791] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.314956] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.315136] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_pool_maxsize = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.315303] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_pool_unused_timeout = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.315469] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_sasl_enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.315721] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_servers = ['localhost:11211'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.315821] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_socket_timeout = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.315982] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.memcache_username = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.316162] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.proxies = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.316331] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.redis_db = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.316496] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.redis_password = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.316666] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.redis_sentinel_service_name = mymaster {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.316885] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.317033] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.redis_server = localhost:6379 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.317213] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.redis_socket_timeout = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.317372] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.redis_username = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.317542] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.retry_attempts = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.317709] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.retry_delay = 0.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.317909] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.socket_keepalive_count = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.318051] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.socket_keepalive_idle = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.318218] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.socket_keepalive_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.318379] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.tls_allowed_ciphers = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.318539] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.tls_cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.318707] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.tls_certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.318860] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.tls_enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.319031] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cache.tls_keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.319213] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.319431] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.auth_type = password {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.319604] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.319786] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.catalog_info = volumev3::publicURL {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.319951] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.320132] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.320324] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.cross_az_attach = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.320526] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.debug = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.320700] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.endpoint_template = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.320882] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.http_retries = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.321065] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.321232] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.321407] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.os_region_name = RegionOne {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.321575] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.321736] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cinder.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.321915] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.322089] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.cpu_dedicated_set = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.322263] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.cpu_shared_set = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.322434] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.image_type_exclude_list = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.322603] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.live_migration_wait_for_vif_plug = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.322770] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.max_concurrent_disk_ops = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.322933] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.max_disk_devices_to_attach = -1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.323118] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.323297] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.323463] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.resource_provider_association_refresh = 300 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.323629] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.323795] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.shutdown_retry_interval = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.323976] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.324175] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] conductor.workers = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.324360] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] console.allowed_origins = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.324525] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] console.ssl_ciphers = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.324699] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] console.ssl_minimum_version = default {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.324869] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] consoleauth.enforce_session_timeout = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.325053] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] consoleauth.token_ttl = 600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.325226] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.325387] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.325554] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.325717] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.connect_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.325879] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.connect_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.326051] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.endpoint_override = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.326218] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.326377] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.326542] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.max_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.326703] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.min_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.326865] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.region_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.327036] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.retriable_status_codes = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.327201] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.service_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.327374] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.service_type = accelerator {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.327541] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.327702] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.status_code_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.327864] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.status_code_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.328038] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.328225] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.328391] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] cyborg.version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.328576] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.backend = sqlalchemy {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.328751] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.connection = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.328922] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.connection_debug = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.329107] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.connection_parameters = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.329308] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.connection_recycle_time = 3600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.329473] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.connection_trace = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.329642] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.db_inc_retry_interval = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.329812] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.db_max_retries = 20 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.329977] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.db_max_retry_interval = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.330157] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.db_retry_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.330343] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.max_overflow = 50 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.330542] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.max_pool_size = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.330714] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.max_retries = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.330889] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.mysql_sql_mode = TRADITIONAL {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.331076] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.mysql_wsrep_sync_wait = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.pool_timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.retry_interval = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.slave_connection = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.sqlite_synchronous = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] database.use_db_reconnect = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.backend = sqlalchemy {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.connection = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332700] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.connection_debug = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332700] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.connection_parameters = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332785] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.connection_recycle_time = 3600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.332899] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.connection_trace = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.333077] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.db_inc_retry_interval = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.333248] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.db_max_retries = 20 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.333415] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.db_max_retry_interval = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.333582] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.db_retry_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.333745] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.max_overflow = 50 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.333909] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.max_pool_size = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.334089] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.max_retries = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.334265] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.334431] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.mysql_wsrep_sync_wait = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.334592] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.pool_timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.334757] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.retry_interval = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.334918] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.slave_connection = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.335092] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] api_database.sqlite_synchronous = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.335274] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] devices.enabled_mdev_types = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.335455] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.335629] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ephemeral_storage_encryption.default_format = luks {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.335795] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ephemeral_storage_encryption.enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.335962] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ephemeral_storage_encryption.key_size = 512 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.336154] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.api_servers = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.336324] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.336489] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.336654] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.336815] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.connect_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.336977] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.connect_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.337159] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.debug = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.337332] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.default_trusted_certificate_ids = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.337494] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.enable_certificate_validation = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.337663] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.enable_rbd_download = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.337820] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.endpoint_override = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.337986] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.338165] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.338325] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.max_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.338485] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.min_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.338651] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.num_retries = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.338837] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.rbd_ceph_conf = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.339083] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.rbd_connect_timeout = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.339284] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.rbd_pool = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.339478] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.rbd_user = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.339650] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.region_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.339816] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.retriable_status_codes = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.339979] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.service_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.340166] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.service_type = image {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.340355] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.340528] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.status_code_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.340692] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.status_code_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.340854] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.341054] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.341228] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.verify_glance_signatures = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.341392] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] glance.version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.341565] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] guestfs.debug = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.341735] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] mks.enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.342113] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.342312] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] image_cache.manager_interval = 2400 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.342486] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] image_cache.precache_concurrency = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.342660] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] image_cache.remove_unused_base_images = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.342834] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.343011] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.343202] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] image_cache.subdirectory_name = _base {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.343384] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.api_max_retries = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.343554] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.api_retry_interval = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.343718] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.343885] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.auth_type = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.344063] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.344226] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.344392] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.344558] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.conductor_group = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.344718] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.connect_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.344878] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.connect_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.345047] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.endpoint_override = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.345213] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.345371] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.345528] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.max_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.345683] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.min_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.345847] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.peer_list = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.346026] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.region_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.346181] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.retriable_status_codes = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.346346] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.serial_console_state_timeout = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.346505] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.service_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.346673] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.service_type = baremetal {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.346832] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.shard = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.346994] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.347171] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.status_code_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.347338] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.status_code_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.347491] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.347671] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.347831] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ironic.version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.348023] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.348204] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] key_manager.fixed_key = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.348389] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.348556] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.barbican_api_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.348720] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.barbican_endpoint = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.348894] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.barbican_endpoint_type = public {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.349068] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.barbican_region_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.349234] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.349416] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.349587] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.349753] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.349913] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.350093] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.number_of_retries = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.350274] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.retry_delay = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.350467] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.send_service_user_token = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.350637] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.350799] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.350963] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.verify_ssl = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.351142] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican.verify_ssl_path = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.351313] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.351481] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.auth_type = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.351652] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.351802] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.351966] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.352142] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.352304] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.352469] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.352632] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] barbican_service_user.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.352799] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.approle_role_id = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.352958] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.approle_secret_id = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.353144] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.kv_mountpoint = secret {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.353308] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.kv_path = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.353477] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.kv_version = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.353637] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.namespace = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.353795] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.root_token_id = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.353954] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.ssl_ca_crt_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.354139] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.timeout = 60.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.354306] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.use_ssl = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.354481] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.354653] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.354819] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.auth_type = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.354981] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.355157] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.355324] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.355484] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.connect_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.355648] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.connect_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.355809] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.endpoint_override = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.355974] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.356149] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.356309] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.max_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.356469] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.min_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.356629] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.region_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.356790] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.retriable_status_codes = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.356949] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.service_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.357133] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.service_type = identity {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.357300] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.357466] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.status_code_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.357628] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.status_code_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.357789] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.357972] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.358149] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] keystone.version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.358355] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.connection_uri = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.358521] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.cpu_mode = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.358689] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.cpu_model_extra_flags = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.358859] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.cpu_models = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.359041] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.cpu_power_governor_high = performance {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.359218] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.cpu_power_governor_low = powersave {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.359413] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.cpu_power_management = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.359593] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.359763] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.device_detach_attempts = 8 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.359928] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.device_detach_timeout = 20 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.360108] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.disk_cachemodes = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.360292] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.disk_prefix = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.360478] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.enabled_perf_events = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.360648] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.file_backed_memory = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.360815] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.gid_maps = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.360979] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.hw_disk_discard = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.361155] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.hw_machine_type = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.361328] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.images_rbd_ceph_conf = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.361498] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.361685] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.361871] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.images_rbd_glance_store_name = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.362059] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.images_rbd_pool = rbd {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.362238] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.images_type = default {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.362404] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.images_volume_group = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.362572] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.inject_key = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.362739] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.inject_partition = -2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.362904] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.inject_password = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.363081] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.iscsi_iface = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.363251] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.iser_use_multipath = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.363418] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_bandwidth = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.363584] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_completion_timeout = 800 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.363749] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_downtime = 500 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.363912] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_downtime_delay = 75 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.364094] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_downtime_steps = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.364261] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_inbound_addr = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.364427] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_permit_auto_converge = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.364589] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_permit_post_copy = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.364749] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_scheme = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.364922] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_timeout_action = abort {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.365101] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_tunnelled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.365270] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_uri = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.365441] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.live_migration_with_native_tls = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.365605] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.max_queues = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.365771] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.mem_stats_period_seconds = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.366021] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.366187] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.nfs_mount_options = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.366876] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.367127] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.num_aoe_discover_tries = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.367330] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.num_iser_scan_tries = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.367504] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.num_memory_encrypted_guests = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.367676] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.num_nvme_discover_tries = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.367846] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.num_pcie_ports = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.368031] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.num_volume_scan_tries = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.368211] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.pmem_namespaces = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.368378] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.quobyte_client_cfg = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.368682] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.368862] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rbd_connect_timeout = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.369045] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.369221] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.369417] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rbd_secret_uuid = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.369589] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rbd_user = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.369758] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.realtime_scheduler_priority = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.369933] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.remote_filesystem_transport = ssh {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.370113] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rescue_image_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.370299] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rescue_kernel_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.370479] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rescue_ramdisk_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.370678] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rng_dev_path = /dev/urandom {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.370853] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.rx_queue_size = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.371129] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.smbfs_mount_options = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.371434] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.371614] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.snapshot_compression = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.371782] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.snapshot_image_format = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.372012] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.372194] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.sparse_logical_volumes = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.372360] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.swtpm_enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.372535] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.swtpm_group = tss {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.372703] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.swtpm_user = tss {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.372872] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.sysinfo_serial = unique {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.373048] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.tb_cache_size = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.373212] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.tx_queue_size = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.373379] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.uid_maps = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.373545] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.use_virtio_for_bridges = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.373718] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.virt_type = kvm {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.373888] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.volume_clear = zero {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.374063] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.volume_clear_size = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.374235] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.volume_use_multipath = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.374396] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.vzstorage_cache_path = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.374568] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.374737] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.vzstorage_mount_group = qemu {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.374901] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.vzstorage_mount_opts = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.375083] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.375365] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.375547] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.vzstorage_mount_user = stack {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.375718] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.375891] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.376077] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.auth_type = password {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.376244] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.376408] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.376577] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.376746] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.connect_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.376904] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.connect_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.377087] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.default_floating_pool = public {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.377254] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.endpoint_override = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.377422] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.extension_sync_interval = 600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.377592] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.http_retries = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.377757] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.377919] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.378091] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.max_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.378266] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.metadata_proxy_shared_secret = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.378432] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.min_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.378604] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.ovs_bridge = br-int {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.378770] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.physnets = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.378941] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.region_name = RegionOne {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.379120] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.retriable_status_codes = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.379325] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.service_metadata_proxy = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.379502] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.service_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.379680] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.service_type = network {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.379853] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.380038] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.status_code_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.380207] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.status_code_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.380373] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.380556] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.380722] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] neutron.version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.380895] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] notifications.bdms_in_notifications = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.381088] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] notifications.default_level = INFO {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.381270] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] notifications.notification_format = unversioned {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.381443] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] notifications.notify_on_state_change = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.381649] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.381895] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] pci.alias = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.382100] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] pci.device_spec = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.382276] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] pci.report_in_placement = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.382457] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.382637] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.auth_type = password {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.382809] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.auth_url = http://10.180.1.21/identity {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.382973] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.383151] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.383317] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.383480] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.connect_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.383644] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.connect_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.383807] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.default_domain_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.383967] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.default_domain_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.384141] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.domain_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.384301] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.domain_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.384463] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.endpoint_override = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.384625] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.384783] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.384941] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.max_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.385112] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.min_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.385284] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.password = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.385446] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.project_domain_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.385615] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.project_domain_name = Default {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.385785] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.project_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.385961] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.project_name = service {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.386148] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.region_name = RegionOne {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.386314] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.retriable_status_codes = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.386480] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.service_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.386651] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.service_type = placement {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.386815] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.386976] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.status_code_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.387161] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.status_code_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.387320] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.system_scope = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.387483] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.387644] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.trust_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.387804] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.user_domain_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.387975] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.user_domain_name = Default {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.388149] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.user_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.388327] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.username = nova {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.388512] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.388675] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] placement.version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.388855] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.cores = 20 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.389034] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.count_usage_from_placement = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.389214] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.389429] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.injected_file_content_bytes = 10240 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.389608] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.injected_file_path_length = 255 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.389779] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.injected_files = 5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.389949] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.instances = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.390137] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.key_pairs = 100 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.390310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.metadata_items = 128 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.390482] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.ram = 51200 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.390650] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.recheck_quota = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.390821] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.server_group_members = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.390988] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] quota.server_groups = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.391182] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.391352] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.391519] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.image_metadata_prefilter = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.391713] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.391892] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.max_attempts = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.392073] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.max_placement_results = 1000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.392243] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.392409] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.query_placement_for_image_type_support = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.392575] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.392753] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] scheduler.workers = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.392927] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.393108] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.393290] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.393464] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.393630] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.393795] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.393959] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.394163] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.394349] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.host_subset_size = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.394504] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.394666] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.image_properties_default_architecture = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.394834] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.395013] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.isolated_hosts = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.395188] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.isolated_images = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.395356] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.max_instances_per_host = 50 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.395521] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.395690] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.395855] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.pci_in_placement = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.396028] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.396201] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.396367] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.396532] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.396697] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.396862] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.397035] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.track_instance_changes = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.397220] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.397392] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] metrics.required = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.397562] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] metrics.weight_multiplier = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.397728] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] metrics.weight_of_unavailable = -10000.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.397893] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] metrics.weight_setting = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.398230] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.398416] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] serial_console.enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.398597] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] serial_console.port_range = 10000:20000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.398773] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.398947] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.399133] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] serial_console.serialproxy_port = 6083 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.399332] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.399528] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.auth_type = password {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.399697] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.399860] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.400039] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.400212] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.400388] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.400581] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.send_service_user_token = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.400753] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.400918] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] service_user.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.401104] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.agent_enabled = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.401274] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.401616] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.401829] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.html5proxy_host = 0.0.0.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.402023] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.html5proxy_port = 6082 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.402193] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.image_compression = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.402358] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.jpeg_compression = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.402521] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.playback_compression = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.402686] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.require_secure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.402856] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.server_listen = 127.0.0.1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.403037] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.403206] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.streaming_mode = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.403366] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] spice.zlib_compression = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.403537] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] upgrade_levels.baseapi = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.403710] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] upgrade_levels.compute = auto {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.403874] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] upgrade_levels.conductor = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.404043] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] upgrade_levels.scheduler = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.404215] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.404384] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.auth_type = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.404556] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.404715] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.404878] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.405050] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.405213] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.405376] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.405539] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vendordata_dynamic_auth.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.405714] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.api_retry_count = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.405877] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.ca_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.406069] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.cache_prefix = devstack-image-cache {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.406300] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.cluster_name = testcl1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.406484] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.connection_pool_size = 10 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.406649] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.console_delay_seconds = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.406822] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.datastore_regex = ^datastore.* {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.407047] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.407233] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.host_password = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.407405] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.host_port = 443 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.407580] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.host_username = administrator@vsphere.local {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.407751] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.insecure = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.407915] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.integration_bridge = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.408095] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.maximum_objects = 100 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.408264] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.pbm_default_policy = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.408433] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.pbm_enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.408594] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.pbm_wsdl_location = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.408763] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.408924] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.serial_port_proxy_uri = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.409095] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.serial_port_service_uri = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.409301] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.task_poll_interval = 0.5 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.409468] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.use_linked_clone = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.409643] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.vnc_keymap = en-us {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.409812] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.vnc_port = 5900 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.409979] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vmware.vnc_port_total = 10000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.410182] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.auth_schemes = ['none'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.410364] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.410669] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.410856] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.411041] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.novncproxy_port = 6080 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.411225] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.server_listen = 127.0.0.1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.411401] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.411567] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.vencrypt_ca_certs = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.411754] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.vencrypt_client_cert = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.411917] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vnc.vencrypt_client_key = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.412114] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.412284] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.disable_deep_image_inspection = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.412453] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.disable_fallback_pcpu_query = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.412618] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.disable_group_policy_check_upcall = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.412782] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.412947] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.disable_rootwrap = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.413124] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.enable_numa_live_migration = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.413294] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.413457] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.413623] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.handle_virt_lifecycle_events = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.413779] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.libvirt_disable_apic = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.413939] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.never_download_image_if_on_rbd = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.414119] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.414290] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.414456] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.414621] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.414785] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.414948] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.415134] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.415301] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.415470] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.415660] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.415834] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.client_socket_timeout = 900 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.416011] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.default_pool_size = 1000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.416195] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.keep_alive = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.416366] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.max_header_line = 16384 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.416532] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.secure_proxy_ssl_header = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.416694] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.ssl_ca_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.416855] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.ssl_cert_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.417024] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.ssl_key_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.417199] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.tcp_keepidle = 600 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.417379] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.417546] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] zvm.ca_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.417711] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] zvm.cloud_connector_url = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.418025] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.418257] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] zvm.reachable_timeout = 300 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.418464] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.enforce_new_defaults = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.418853] env[69171]: WARNING oslo_config.cfg [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 618.419062] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.enforce_scope = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.419253] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.policy_default_rule = default {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.419470] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.419655] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.policy_file = policy.yaml {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.419836] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.420008] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.420185] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.420363] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.420528] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.420701] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.420877] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.421067] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.connection_string = messaging:// {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.421243] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.enabled = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.421416] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.es_doc_type = notification {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.421592] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.es_scroll_size = 10000 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.421789] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.es_scroll_time = 2m {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.421959] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.filter_error_trace = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.422145] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.hmac_keys = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.422319] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.sentinel_service_name = mymaster {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.422494] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.socket_timeout = 0.1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.422662] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.trace_requests = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.422825] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler.trace_sqlalchemy = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.423009] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler_jaeger.process_tags = {} {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.423183] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler_jaeger.service_name_prefix = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.423349] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] profiler_otlp.service_name_prefix = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.423520] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] remote_debug.host = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.423696] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] remote_debug.port = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.423863] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.424040] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.424213] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.424380] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.424545] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.424710] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.424876] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.425051] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.425220] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.425391] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.hostname = devstack {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.425554] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.425725] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.425893] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.426076] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.426251] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.426420] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.426588] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.426763] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.426928] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.427105] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.427279] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.427445] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.427610] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.427777] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.427940] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.428116] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.428282] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.428443] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.428611] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.428780] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.ssl = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.428955] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.429138] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.429330] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.429511] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.429687] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.ssl_version = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.429853] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.430056] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.430232] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_notifications.retry = -1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.430470] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.430665] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_messaging_notifications.transport_url = **** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.430843] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.auth_section = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.431021] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.auth_type = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.431214] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.cafile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.431428] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.certfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.431606] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.collect_timing = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.431770] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.connect_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.431934] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.connect_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.432109] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.endpoint_id = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.432272] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.endpoint_override = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.432484] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.insecure = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.432662] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.keyfile = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.432825] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.max_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.432985] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.min_version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.433163] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.region_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.433332] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.retriable_status_codes = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.433496] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.service_name = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.433657] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.service_type = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.433821] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.split_loggers = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.433983] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.status_code_retries = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.434160] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.status_code_retry_delay = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.434480] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.timeout = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.434530] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.valid_interfaces = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.434679] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_limit.version = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.434852] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_reports.file_event_handler = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.435039] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_reports.file_event_handler_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.435211] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] oslo_reports.log_dir = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.435388] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.435553] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_linux_bridge_privileged.group = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.435713] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.435881] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.436062] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.436230] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_linux_bridge_privileged.user = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.436404] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.436567] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_ovs_privileged.group = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.436726] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_ovs_privileged.helper_command = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.436894] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.437073] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.437254] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] vif_plug_ovs_privileged.user = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.437413] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.flat_interface = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.437596] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.437771] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.437946] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.438133] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.438310] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.438481] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.438645] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_linux_bridge.vlan_interface = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.438825] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.438997] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_ovs.isolate_vif = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.439188] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.439395] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.439571] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.439744] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_ovs.ovsdb_interface = native {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.439909] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_vif_ovs.per_port_bridge = False {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.440090] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_brick.lock_path = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.440271] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.440457] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] os_brick.wait_mpath_device_interval = 1 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.440633] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] privsep_osbrick.capabilities = [21] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.440795] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] privsep_osbrick.group = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.440954] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] privsep_osbrick.helper_command = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.441137] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.441307] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] privsep_osbrick.thread_pool_size = 8 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.441470] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] privsep_osbrick.user = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.441644] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.441804] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] nova_sys_admin.group = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.441964] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] nova_sys_admin.helper_command = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.442145] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.442314] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] nova_sys_admin.thread_pool_size = 8 {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.442476] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] nova_sys_admin.user = None {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 618.442608] env[69171]: DEBUG oslo_service.service [None req-3ba9097a-78e6-4a10-bcd7-238c3a6d09fc None None] ******************************************************************************** {{(pid=69171) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 618.443111] env[69171]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 618.946820] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Getting list of instances from cluster (obj){ [ 618.946820] env[69171]: value = "domain-c8" [ 618.946820] env[69171]: _type = "ClusterComputeResource" [ 618.946820] env[69171]: } {{(pid=69171) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 618.948067] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cefb4fb0-266f-4a99-9f75-5047d940988a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.957435] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Got total of 0 instances {{(pid=69171) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 618.958013] env[69171]: WARNING nova.virt.vmwareapi.driver [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 618.958546] env[69171]: INFO nova.virt.node [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Generated node identity 560320dd-0035-4785-a81a-5134ce1f7aaa [ 618.958789] env[69171]: INFO nova.virt.node [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Wrote node identity 560320dd-0035-4785-a81a-5134ce1f7aaa to /opt/stack/data/n-cpu-1/compute_id [ 619.461625] env[69171]: WARNING nova.compute.manager [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Compute nodes ['560320dd-0035-4785-a81a-5134ce1f7aaa'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 620.467440] env[69171]: INFO nova.compute.manager [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 621.473717] env[69171]: WARNING nova.compute.manager [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 621.474136] env[69171]: DEBUG oslo_concurrency.lockutils [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.474232] env[69171]: DEBUG oslo_concurrency.lockutils [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.474378] env[69171]: DEBUG oslo_concurrency.lockutils [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.474560] env[69171]: DEBUG nova.compute.resource_tracker [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=69171) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 621.475538] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba50c2d-9502-40da-8ebb-eb5a21ccf6d9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.485296] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08aad7cf-0ad1-4274-a7bf-6a5a49efebcd {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.500193] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-989a9fab-eeaf-4559-90a6-339d0dbadcb8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.508050] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c5be99-3469-494d-abc9-68bd91b0ae5a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.537493] env[69171]: DEBUG nova.compute.resource_tracker [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180732MB free_disk=1GB free_vcpus=48 pci_devices=None {{(pid=69171) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 621.537659] env[69171]: DEBUG oslo_concurrency.lockutils [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.537852] env[69171]: DEBUG oslo_concurrency.lockutils [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.040569] env[69171]: WARNING nova.compute.resource_tracker [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] No compute node record for cpu-1:560320dd-0035-4785-a81a-5134ce1f7aaa: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 560320dd-0035-4785-a81a-5134ce1f7aaa could not be found. [ 622.544144] env[69171]: INFO nova.compute.resource_tracker [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 560320dd-0035-4785-a81a-5134ce1f7aaa [ 624.052722] env[69171]: DEBUG nova.compute.resource_tracker [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 624.053101] env[69171]: DEBUG nova.compute.resource_tracker [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 624.219599] env[69171]: INFO nova.scheduler.client.report [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] [req-f74615fa-ee68-4dca-80d2-dd2926fea088] Created resource provider record via placement API for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 624.235961] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4a938b7-6815-4569-8b85-e7811db02e69 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.245068] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf7991a-7443-4821-9571-f149c384e208 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.274707] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57498b2e-3f1d-41d0-afee-e759aad84d0e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.282662] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44f5dce-ea9a-4764-a026-c3dd2ec5f3ac {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.296558] env[69171]: DEBUG nova.compute.provider_tree [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 624.834933] env[69171]: DEBUG nova.scheduler.client.report [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Updated inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 624.835191] env[69171]: DEBUG nova.compute.provider_tree [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Updating resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa generation from 0 to 1 during operation: update_inventory {{(pid=69171) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 624.835336] env[69171]: DEBUG nova.compute.provider_tree [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 624.885726] env[69171]: DEBUG nova.compute.provider_tree [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Updating resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa generation from 1 to 2 during operation: update_traits {{(pid=69171) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 625.391022] env[69171]: DEBUG nova.compute.resource_tracker [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=69171) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 625.391472] env[69171]: DEBUG oslo_concurrency.lockutils [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.853s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.391472] env[69171]: DEBUG nova.service [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Creating RPC server for service compute {{(pid=69171) start /opt/stack/nova/nova/service.py:186}} [ 625.411744] env[69171]: DEBUG nova.service [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] Join ServiceGroup membership for this service compute {{(pid=69171) start /opt/stack/nova/nova/service.py:203}} [ 625.411992] env[69171]: DEBUG nova.servicegroup.drivers.db [None req-602a4f0c-ad9b-4a5f-9a6e-13b01ea99b09 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=69171) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 628.413779] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._sync_power_states {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 628.918551] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Getting list of instances from cluster (obj){ [ 628.918551] env[69171]: value = "domain-c8" [ 628.918551] env[69171]: _type = "ClusterComputeResource" [ 628.918551] env[69171]: } {{(pid=69171) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 628.919811] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1141309-9301-45fd-8aea-b83de8c055c8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.929080] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Got total of 0 instances {{(pid=69171) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 628.929328] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 628.929688] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Getting list of instances from cluster (obj){ [ 628.929688] env[69171]: value = "domain-c8" [ 628.929688] env[69171]: _type = "ClusterComputeResource" [ 628.929688] env[69171]: } {{(pid=69171) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2122}} [ 628.930728] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd76e3f5-3b91-469b-85d5-454ce7dfa58c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.938708] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Got total of 0 instances {{(pid=69171) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2131}} [ 656.707470] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "f8b20421-15d3-46f3-a504-172d044ff4d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.707746] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.712559] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.712806] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.211215] env[69171]: DEBUG nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 657.218476] env[69171]: DEBUG nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 657.750698] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.751117] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.752549] env[69171]: INFO nova.compute.claims [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.760487] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.846929] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5698284-a7b3-464d-834f-8b58338f42e8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.858805] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9c7d79f-416e-4f08-a378-521bd5004423 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.904985] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353dd32a-714a-409d-befa-0faaa85e87b4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.914029] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e1fbed-4397-4bdd-a58d-dae01a6f1ca0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.929678] env[69171]: DEBUG nova.compute.provider_tree [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Inventory has not changed in ProviderTree for provider: 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 659.433508] env[69171]: DEBUG nova.scheduler.client.report [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Inventory has not changed for provider 560320dd-0035-4785-a81a-5134ce1f7aaa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.944267] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.193s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.945464] env[69171]: DEBUG nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Start building networks asynchronously for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2830}} [ 659.949080] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.188s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.951595] env[69171]: INFO nova.compute.claims [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 660.457350] env[69171]: DEBUG nova.compute.utils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Using /dev/sd instead of None {{(pid=69171) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.460273] env[69171]: DEBUG nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Allocating IP information in the background. {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1982}} [ 660.462532] env[69171]: DEBUG nova.network.neutron [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] allocate_for_instance() {{(pid=69171) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 660.963022] env[69171]: DEBUG nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Start building block device mappings for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2865}} [ 661.058644] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea0ab38-8a1c-4155-8ffa-7d085997c373 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.070052] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad41567-c66e-4771-b915-a4d8477e2da5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.114287] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050e388d-9d79-49e0-8574-a663b6b61e90 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.122773] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2b530fc-f037-4eb5-84ad-20e98dcd2587 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.148485] env[69171]: DEBUG nova.compute.provider_tree [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Inventory has not changed in ProviderTree for provider: 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 661.649800] env[69171]: DEBUG nova.scheduler.client.report [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Inventory has not changed for provider 560320dd-0035-4785-a81a-5134ce1f7aaa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.972869] env[69171]: DEBUG nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Start spawning the instance on the hypervisor. {{(pid=69171) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2639}} [ 662.159509] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.207s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.159509] env[69171]: DEBUG nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Start building networks asynchronously for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2830}} [ 662.667437] env[69171]: DEBUG nova.compute.utils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Using /dev/sd instead of None {{(pid=69171) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.669310] env[69171]: DEBUG nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Allocating IP information in the background. {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1982}} [ 662.669771] env[69171]: DEBUG nova.network.neutron [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] allocate_for_instance() {{(pid=69171) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 662.823486] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-22T21:48:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-22T21:48:14Z,direct_url=,disk_format='vmdk',id=8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1f8548ca1c9947769d22274768a59ff7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-22T21:48:15Z,virtual_size=,visibility=), allow threads: False {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 662.823771] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Flavor limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 662.823881] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Image limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 662.827652] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Flavor pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 662.828037] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Image pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 662.830732] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 662.830732] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 662.830732] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 662.830732] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Got 1 possible topologies {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 662.830732] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 662.831123] env[69171]: DEBUG nova.virt.hardware [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 662.832154] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a40eb1a3-b58e-464c-8351-48fe7facd75d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.842044] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78bf5d3-f40c-4823-965c-2ab43d88031a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.881651] env[69171]: DEBUG nova.policy [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb5d5af0cd4449c4b577c02bf053f571', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c585337456cb43cd88b98586d3d9eea1', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=69171) authorize /opt/stack/nova/nova/policy.py:201}} [ 662.888685] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d87e09-9559-43a6-beec-62a197d50b2c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.936903] env[69171]: DEBUG nova.policy [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b3d9fb8d4ba4df393db8e93ec169b86', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '914e6c234dec4422a7448f447f80b588', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=69171) authorize /opt/stack/nova/nova/policy.py:201}} [ 663.170556] env[69171]: DEBUG nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Start building block device mappings for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2865}} [ 663.397540] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "8e6a9d73-6705-484c-82e0-70c6a6860211" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.397793] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.901696] env[69171]: DEBUG nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 664.186192] env[69171]: DEBUG nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Start spawning the instance on the hypervisor. {{(pid=69171) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2639}} [ 664.236503] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-22T21:48:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-22T21:48:14Z,direct_url=,disk_format='vmdk',id=8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1f8548ca1c9947769d22274768a59ff7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-22T21:48:15Z,virtual_size=,visibility=), allow threads: False {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 664.236503] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Flavor limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 664.236802] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Image limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 664.238409] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Flavor pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 664.238409] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Image pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 664.238409] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 664.238409] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 664.238409] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 664.238605] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Got 1 possible topologies {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 664.238605] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 664.241104] env[69171]: DEBUG nova.virt.hardware [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 664.241104] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6c1efe3-ca3d-45af-9b83-3252026b7fd9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.250989] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca57f025-3d32-4a77-a708-fb4e4d02a7a2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.339098] env[69171]: DEBUG nova.network.neutron [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Successfully created port: aad071eb-62cd-4e3d-b8fc-62ea5e458141 {{(pid=69171) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 664.449033] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 664.449925] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.452430] env[69171]: INFO nova.compute.claims [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.901609] env[69171]: DEBUG nova.network.neutron [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Successfully created port: 8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 {{(pid=69171) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.562101] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d95fe0d-9f4b-48d6-bd97-7d411a50b92d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.571247] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef00cce-ed15-44dd-8a98-f69f3cd92703 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.611524] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e82bbd2-f0a4-44b5-aa8f-6f4c80f3de74 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.627031] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb8ca38c-9e45-443c-ae42-4e79aec58a8b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.644067] env[69171]: DEBUG nova.compute.provider_tree [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Inventory has not changed in ProviderTree for provider: 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 666.148044] env[69171]: DEBUG nova.scheduler.client.report [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Inventory has not changed for provider 560320dd-0035-4785-a81a-5134ce1f7aaa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.160448] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "6a29d495-c938-41ad-97b6-1331b6eff589" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.161401] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "6a29d495-c938-41ad-97b6-1331b6eff589" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.657319] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.205s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.658075] env[69171]: DEBUG nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Start building networks asynchronously for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2830}} [ 666.665993] env[69171]: DEBUG nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 666.924687] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 666.925040] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.144370] env[69171]: DEBUG nova.network.neutron [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Successfully updated port: aad071eb-62cd-4e3d-b8fc-62ea5e458141 {{(pid=69171) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 667.170042] env[69171]: DEBUG nova.compute.utils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Using /dev/sd instead of None {{(pid=69171) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.177719] env[69171]: DEBUG nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Allocating IP information in the background. {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1982}} [ 667.177719] env[69171]: DEBUG nova.network.neutron [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] allocate_for_instance() {{(pid=69171) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 667.216378] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.216612] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 667.218447] env[69171]: INFO nova.compute.claims [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 667.300863] env[69171]: DEBUG nova.policy [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cfade167d4024c2193395b1a069306d9', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b3c1c9908738493e8b1a6c0d90b59b05', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=69171) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.431661] env[69171]: DEBUG nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 667.651665] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.651665] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquired lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.651665] env[69171]: DEBUG nova.network.neutron [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 667.677613] env[69171]: DEBUG nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Start building block device mappings for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2865}} [ 667.967048] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.007170] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.007359] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.063201] env[69171]: DEBUG nova.network.neutron [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Successfully created port: 547da9e0-630e-4393-95ef-f0b0086a5b94 {{(pid=69171) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 668.322268] env[69171]: DEBUG nova.network.neutron [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 668.355519] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01ef000f-c0ca-4a4b-bfab-589b02bc85dc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.363490] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdeb582d-6abe-403b-8905-202e5a914f71 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.407518] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03f744bf-2428-4096-96e6-6e0cb7423f99 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.417195] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-829ad901-c6cc-42b9-b3bb-25338cf716f2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.437699] env[69171]: DEBUG nova.compute.provider_tree [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Inventory has not changed in ProviderTree for provider: 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 668.512367] env[69171]: DEBUG nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 668.627398] env[69171]: DEBUG nova.network.neutron [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Successfully updated port: 8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 {{(pid=69171) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 668.690768] env[69171]: DEBUG nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Start spawning the instance on the hypervisor. {{(pid=69171) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2639}} [ 668.736436] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-22T21:48:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-22T21:48:14Z,direct_url=,disk_format='vmdk',id=8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1f8548ca1c9947769d22274768a59ff7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-22T21:48:15Z,virtual_size=,visibility=), allow threads: False {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.736678] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Flavor limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.736813] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Image limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.737154] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Flavor pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.737343] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Image pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.737462] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.737674] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.737846] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.738066] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Got 1 possible topologies {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.738238] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.738415] env[69171]: DEBUG nova.virt.hardware [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.739310] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7889149e-0fb2-4be4-91b7-9537bd644958 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.749706] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6cdb9f7-92b6-4ef3-9db4-4303cf0779d7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.901820] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Acquiring lock "f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 668.902279] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Lock "f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.942997] env[69171]: DEBUG nova.scheduler.client.report [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Inventory has not changed for provider 560320dd-0035-4785-a81a-5134ce1f7aaa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 669.053831] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.130832] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 669.131052] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquired lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 669.131292] env[69171]: DEBUG nova.network.neutron [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 669.173985] env[69171]: DEBUG nova.network.neutron [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Updating instance_info_cache with network_info: [{"id": "aad071eb-62cd-4e3d-b8fc-62ea5e458141", "address": "fa:16:3e:d7:75:e3", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaad071eb-62", "ovs_interfaceid": "aad071eb-62cd-4e3d-b8fc-62ea5e458141", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.406583] env[69171]: DEBUG nova.compute.manager [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 669.455394] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.239s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 669.458403] env[69171]: DEBUG nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Start building networks asynchronously for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2830}} [ 669.462156] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.498s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 669.464598] env[69171]: INFO nova.compute.claims [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 669.679610] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Releasing lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.680130] env[69171]: DEBUG nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Instance network_info: |[{"id": "aad071eb-62cd-4e3d-b8fc-62ea5e458141", "address": "fa:16:3e:d7:75:e3", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaad071eb-62", "ovs_interfaceid": "aad071eb-62cd-4e3d-b8fc-62ea5e458141", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1997}} [ 669.680664] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d7:75:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'aad071eb-62cd-4e3d-b8fc-62ea5e458141', 'vif_model': 'vmxnet3'}] {{(pid=69171) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 669.694911] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 669.695627] env[69171]: DEBUG nova.network.neutron [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 669.702795] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-736426cf-ce7e-4d2c-9254-2f72981b89a7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.717065] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Created folder: OpenStack in parent group-v4. [ 669.717471] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Creating folder: Project (c585337456cb43cd88b98586d3d9eea1). Parent ref: group-v838162. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 669.718565] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0680a990-1b37-4b50-8ef6-5df2e6d2424c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.734028] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Created folder: Project (c585337456cb43cd88b98586d3d9eea1) in parent group-v838162. [ 669.734028] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Creating folder: Instances. Parent ref: group-v838163. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 669.734028] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-96e2e1f4-df92-470d-ace8-5fbfce7a4ac7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.745642] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Created folder: Instances in parent group-v838163. [ 669.746437] env[69171]: DEBUG oslo.service.loopingcall [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 669.746841] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Creating VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 669.747234] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3920be32-0e71-4bd9-b47c-69e4dbd654a3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.770391] env[69171]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 669.770391] env[69171]: value = "task-4235862" [ 669.770391] env[69171]: _type = "Task" [ 669.770391] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 669.778690] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235862, 'name': CreateVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 669.946714] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.975643] env[69171]: DEBUG nova.compute.utils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Using /dev/sd instead of None {{(pid=69171) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 669.979023] env[69171]: DEBUG nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Not allocating networking since 'none' was specified. {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1978}} [ 670.027314] env[69171]: DEBUG nova.network.neutron [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updating instance_info_cache with network_info: [{"id": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "address": "fa:16:3e:83:7a:e9", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cd0dd84-ce", "ovs_interfaceid": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.284351] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235862, 'name': CreateVM_Task, 'duration_secs': 0.459756} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 670.284602] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Created VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 670.295327] env[69171]: DEBUG oslo_vmware.service [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a80dcd4-4099-4e51-ba23-1a3d2ff0e299 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.309083] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.309083] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.309083] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 670.310075] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4c66d9f-612f-48f7-bd62-0c563234916d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.319079] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 670.319079] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52f6140e-3ad6-c8e9-a612-7c32a086a7b7" [ 670.319079] env[69171]: _type = "Task" [ 670.319079] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.326713] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52f6140e-3ad6-c8e9-a612-7c32a086a7b7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.459578] env[69171]: DEBUG nova.compute.manager [req-ad17d2e2-b6cb-4da9-bf49-f44dcc16749a req-0839c56d-7a30-4ffa-9992-b60cebd409ef service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Received event network-vif-plugged-aad071eb-62cd-4e3d-b8fc-62ea5e458141 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 670.459813] env[69171]: DEBUG oslo_concurrency.lockutils [req-ad17d2e2-b6cb-4da9-bf49-f44dcc16749a req-0839c56d-7a30-4ffa-9992-b60cebd409ef service nova] Acquiring lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.460086] env[69171]: DEBUG oslo_concurrency.lockutils [req-ad17d2e2-b6cb-4da9-bf49-f44dcc16749a req-0839c56d-7a30-4ffa-9992-b60cebd409ef service nova] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.460270] env[69171]: DEBUG oslo_concurrency.lockutils [req-ad17d2e2-b6cb-4da9-bf49-f44dcc16749a req-0839c56d-7a30-4ffa-9992-b60cebd409ef service nova] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.460441] env[69171]: DEBUG nova.compute.manager [req-ad17d2e2-b6cb-4da9-bf49-f44dcc16749a req-0839c56d-7a30-4ffa-9992-b60cebd409ef service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] No waiting events found dispatching network-vif-plugged-aad071eb-62cd-4e3d-b8fc-62ea5e458141 {{(pid=69171) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 670.460699] env[69171]: WARNING nova.compute.manager [req-ad17d2e2-b6cb-4da9-bf49-f44dcc16749a req-0839c56d-7a30-4ffa-9992-b60cebd409ef service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Received unexpected event network-vif-plugged-aad071eb-62cd-4e3d-b8fc-62ea5e458141 for instance with vm_state building and task_state spawning. [ 670.483282] env[69171]: DEBUG nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Start building block device mappings for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2865}} [ 670.533170] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Releasing lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.539428] env[69171]: DEBUG nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Instance network_info: |[{"id": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "address": "fa:16:3e:83:7a:e9", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cd0dd84-ce", "ovs_interfaceid": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1997}} [ 670.539735] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:83:7a:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7', 'vif_model': 'vmxnet3'}] {{(pid=69171) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 670.554023] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Creating folder: Project (914e6c234dec4422a7448f447f80b588). Parent ref: group-v838162. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 670.554283] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-374113c3-5411-46a9-8372-97b80f871810 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.571365] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Created folder: Project (914e6c234dec4422a7448f447f80b588) in parent group-v838162. [ 670.571365] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Creating folder: Instances. Parent ref: group-v838166. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 670.571365] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-85239906-8b30-4823-a2a7-94c4c9840bc1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.582795] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Created folder: Instances in parent group-v838166. [ 670.583033] env[69171]: DEBUG oslo.service.loopingcall [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.583160] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Creating VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 670.583610] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4fc9cf76-a11c-4fba-8285-74bbc3742f16 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.612465] env[69171]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 670.612465] env[69171]: value = "task-4235865" [ 670.612465] env[69171]: _type = "Task" [ 670.612465] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.624996] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235865, 'name': CreateVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.683332] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8b1648e-c415-4730-8f6e-f7874790cc2d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.703496] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54d8143-f223-43ce-8362-ae06c621ea34 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.743695] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3ecae3-7a27-4858-9110-e67a703a0c79 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.755629] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc4c2db-97ce-4604-97e1-edc5fd3f6846 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.770999] env[69171]: DEBUG nova.compute.provider_tree [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Inventory has not changed in ProviderTree for provider: 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.805672] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "23e40e0a-45e2-4735-b953-2c165d318c50" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.805672] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "23e40e0a-45e2-4735-b953-2c165d318c50" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.834566] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.835619] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Processing image 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 670.835619] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.835619] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.835936] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 670.837042] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a2128204-4029-4f66-960b-aac59f0a78ee {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.860334] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 670.860553] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=69171) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 670.861489] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcb41db6-3b9e-43e7-b6d7-db922e35c518 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.870812] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d19c9b37-ca73-4580-aa01-1d4e7494e5d5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.880215] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 670.880215] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]527754e5-cbab-d3ae-584d-d4989f554ac5" [ 670.880215] env[69171]: _type = "Task" [ 670.880215] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 670.891323] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]527754e5-cbab-d3ae-584d-d4989f554ac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 670.894361] env[69171]: DEBUG nova.compute.manager [req-2e558c59-a766-4c0b-8d66-4daf2cc14415 req-c4b8990b-8333-4a7e-ba29-0f9d6776650e service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Received event network-vif-plugged-8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 670.894599] env[69171]: DEBUG oslo_concurrency.lockutils [req-2e558c59-a766-4c0b-8d66-4daf2cc14415 req-c4b8990b-8333-4a7e-ba29-0f9d6776650e service nova] Acquiring lock "f8b20421-15d3-46f3-a504-172d044ff4d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.895168] env[69171]: DEBUG oslo_concurrency.lockutils [req-2e558c59-a766-4c0b-8d66-4daf2cc14415 req-c4b8990b-8333-4a7e-ba29-0f9d6776650e service nova] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.895367] env[69171]: DEBUG oslo_concurrency.lockutils [req-2e558c59-a766-4c0b-8d66-4daf2cc14415 req-c4b8990b-8333-4a7e-ba29-0f9d6776650e service nova] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.895533] env[69171]: DEBUG nova.compute.manager [req-2e558c59-a766-4c0b-8d66-4daf2cc14415 req-c4b8990b-8333-4a7e-ba29-0f9d6776650e service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] No waiting events found dispatching network-vif-plugged-8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 {{(pid=69171) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 670.895736] env[69171]: WARNING nova.compute.manager [req-2e558c59-a766-4c0b-8d66-4daf2cc14415 req-c4b8990b-8333-4a7e-ba29-0f9d6776650e service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Received unexpected event network-vif-plugged-8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 for instance with vm_state building and task_state spawning. [ 670.987556] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Acquiring lock "836c5a40-0325-49d5-90d2-bd86cb3748ab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 670.987793] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Lock "836c5a40-0325-49d5-90d2-bd86cb3748ab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.125776] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235865, 'name': CreateVM_Task, 'duration_secs': 0.400134} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 671.126056] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Created VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 671.126816] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.126979] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.127373] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 671.127599] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-136aab32-6204-4b4e-881a-e4f3a983759c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.133908] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 671.133908] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52cb6266-72a8-5ed4-c1ca-691c757e50a6" [ 671.133908] env[69171]: _type = "Task" [ 671.133908] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.144177] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52cb6266-72a8-5ed4-c1ca-691c757e50a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.274101] env[69171]: DEBUG nova.scheduler.client.report [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Inventory has not changed for provider 560320dd-0035-4785-a81a-5134ce1f7aaa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.307761] env[69171]: DEBUG nova.compute.manager [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 671.389760] env[69171]: DEBUG nova.network.neutron [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Successfully updated port: 547da9e0-630e-4393-95ef-f0b0086a5b94 {{(pid=69171) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 671.395178] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Preparing fetch location {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 671.396426] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Creating directory with path [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 671.396426] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4426dcde-4ef4-4a9c-93b8-698410776361 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.445169] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Created directory with path [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 671.445169] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Fetch image to [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 671.445321] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Downloading image file data 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 to [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk on the data store datastore1 {{(pid=69171) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 671.446145] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dff41960-7527-4089-91fa-a832c3cd42a7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.456142] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1574c8b6-7bac-4bb9-a1b6-f6a582e55ec2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.470052] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740977e3-63ef-4297-bdf0-bbcf0a6b2e78 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.507591] env[69171]: DEBUG nova.compute.manager [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 671.514851] env[69171]: DEBUG nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Start spawning the instance on the hypervisor. {{(pid=69171) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2639}} [ 671.519898] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a9e0504-36e4-4e0c-8cbc-e4ee0396973c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.526718] env[69171]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-a3ff93bd-2a54-4081-8dca-8c9d54310cc0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.550994] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-22T21:48:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-22T21:48:14Z,direct_url=,disk_format='vmdk',id=8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1f8548ca1c9947769d22274768a59ff7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-22T21:48:15Z,virtual_size=,visibility=), allow threads: False {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 671.551274] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Flavor limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 671.551411] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Image limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 671.551623] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Flavor pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 671.551740] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Image pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 671.551871] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 671.552647] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 671.552647] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 671.552647] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Got 1 possible topologies {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 671.552647] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 671.552921] env[69171]: DEBUG nova.virt.hardware [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 671.553683] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65b057e0-9c45-428e-8535-f1acfecd3005 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.574981] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff3a693-7ca7-45ea-9197-9e0b853054ae {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.594266] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Instance VIF info [] {{(pid=69171) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 671.600013] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Creating folder: Project (708d12d06a4149fcbcecf9be6d6ec288). Parent ref: group-v838162. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 671.600721] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2d87d75e-d5eb-4159-a008-0586bb6cc04f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.612362] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Created folder: Project (708d12d06a4149fcbcecf9be6d6ec288) in parent group-v838162. [ 671.612633] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Creating folder: Instances. Parent ref: group-v838169. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 671.612957] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-507193b1-7857-46c2-ac76-51a89a534d0f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.623791] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Created folder: Instances in parent group-v838169. [ 671.624229] env[69171]: DEBUG oslo.service.loopingcall [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 671.624457] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Creating VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 671.624912] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8c4cc275-b96b-4743-9c3b-fdf4bc718104 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.645995] env[69171]: DEBUG nova.virt.vmwareapi.images [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Downloading image file data 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 to the data store datastore1 {{(pid=69171) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 671.663200] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.663200] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Processing image 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 671.663200] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.663200] env[69171]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 671.663200] env[69171]: value = "task-4235868" [ 671.663200] env[69171]: _type = "Task" [ 671.663200] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 671.672418] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235868, 'name': CreateVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 671.740073] env[69171]: DEBUG oslo_vmware.rw_handles [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=69171) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 671.811038] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.349s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.811698] env[69171]: DEBUG nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Start building networks asynchronously for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2830}} [ 671.818501] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.764s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 671.822946] env[69171]: INFO nova.compute.claims [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 671.875651] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.902558] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.902711] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquired lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.902864] env[69171]: DEBUG nova.network.neutron [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 672.051080] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 672.182421] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235868, 'name': CreateVM_Task, 'duration_secs': 0.387771} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 672.185473] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Created VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 672.187768] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.187961] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.188330] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 672.188672] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b33c11a1-8891-4e6b-b283-9266a344f2c7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.198124] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 672.198124] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]529887fb-9edd-d9c1-db43-481f5afd5c4a" [ 672.198124] env[69171]: _type = "Task" [ 672.198124] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.226873] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.227285] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Processing image 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 672.227434] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.342916] env[69171]: DEBUG nova.compute.utils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Using /dev/sd instead of None {{(pid=69171) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 672.348713] env[69171]: DEBUG nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Allocating IP information in the background. {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1982}} [ 672.348713] env[69171]: DEBUG nova.network.neutron [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] allocate_for_instance() {{(pid=69171) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 672.530780] env[69171]: DEBUG nova.network.neutron [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 672.591159] env[69171]: DEBUG oslo_vmware.rw_handles [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Completed reading data from the image iterator. {{(pid=69171) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 672.591398] env[69171]: DEBUG oslo_vmware.rw_handles [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Closing write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=69171) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 672.595418] env[69171]: DEBUG nova.policy [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'aa46788c1ed041b49f1e197394cbe784', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f851949057dc4a268959f6c7cf0bd632', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=69171) authorize /opt/stack/nova/nova/policy.py:201}} [ 672.675824] env[69171]: DEBUG nova.virt.vmwareapi.images [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Downloaded image file data 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 to vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk on the data store datastore1 {{(pid=69171) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 672.679024] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Caching image {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 672.679024] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Copying Virtual Disk [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk to [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 672.679513] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-77a03eec-8673-4205-8d80-800c0cfbb62c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.690018] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 672.690018] env[69171]: value = "task-4235869" [ 672.690018] env[69171]: _type = "Task" [ 672.690018] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 672.699440] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235869, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 672.849126] env[69171]: DEBUG nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Start building block device mappings for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2865}} [ 673.038213] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a569abf-86f0-44f3-9121-8a62b9d304b9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.048448] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41018a42-fd4f-4183-8782-9a6c3b465df5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.053473] env[69171]: DEBUG nova.network.neutron [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Updating instance_info_cache with network_info: [{"id": "547da9e0-630e-4393-95ef-f0b0086a5b94", "address": "fa:16:3e:4f:61:19", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap547da9e0-63", "ovs_interfaceid": "547da9e0-630e-4393-95ef-f0b0086a5b94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.088435] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33de7778-ec81-4390-ad59-df061611b10e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.098608] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a3c36a4-e632-4978-a9d3-c3d96c53520f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.115053] env[69171]: DEBUG nova.compute.provider_tree [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Inventory has not changed in ProviderTree for provider: 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 673.204517] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235869, 'name': CopyVirtualDisk_Task} progress is 27%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.331888] env[69171]: DEBUG nova.network.neutron [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Successfully created port: d8a791ab-d322-465f-80d1-73b383b7205b {{(pid=69171) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 673.555796] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Releasing lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.556551] env[69171]: DEBUG nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Instance network_info: |[{"id": "547da9e0-630e-4393-95ef-f0b0086a5b94", "address": "fa:16:3e:4f:61:19", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap547da9e0-63", "ovs_interfaceid": "547da9e0-630e-4393-95ef-f0b0086a5b94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1997}} [ 673.556758] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4f:61:19', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '547da9e0-630e-4393-95ef-f0b0086a5b94', 'vif_model': 'vmxnet3'}] {{(pid=69171) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 673.567105] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Creating folder: Project (b3c1c9908738493e8b1a6c0d90b59b05). Parent ref: group-v838162. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 673.567454] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-56c884d9-ad5c-4af9-a09d-31baaf53f4e1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.579176] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Created folder: Project (b3c1c9908738493e8b1a6c0d90b59b05) in parent group-v838162. [ 673.580371] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Creating folder: Instances. Parent ref: group-v838172. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 673.580371] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9e145b6b-a456-4230-b779-c80ff66543c8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.590024] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.590291] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 673.590562] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Starting heal instance info cache {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9929}} [ 673.590654] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Rebuilding the list of instances to heal {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9933}} [ 673.594184] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Created folder: Instances in parent group-v838172. [ 673.594558] env[69171]: DEBUG oslo.service.loopingcall [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 673.595098] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Creating VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 673.595382] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-506e5147-1807-4e4a-8c74-152a43f5d4a4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.620113] env[69171]: DEBUG nova.scheduler.client.report [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Inventory has not changed for provider 560320dd-0035-4785-a81a-5134ce1f7aaa based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.624151] env[69171]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 673.624151] env[69171]: value = "task-4235872" [ 673.624151] env[69171]: _type = "Task" [ 673.624151] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.635871] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235872, 'name': CreateVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.702679] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235869, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.943832} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 673.703424] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Copied Virtual Disk [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk to [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 673.703806] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Deleting the datastore file [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/tmp-sparse.vmdk {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 673.704182] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c82895c8-3b5c-4a6f-9924-7eea7dc2401d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.711981] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 673.711981] env[69171]: value = "task-4235873" [ 673.711981] env[69171]: _type = "Task" [ 673.711981] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 673.722094] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235873, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 673.867787] env[69171]: DEBUG nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Start spawning the instance on the hypervisor. {{(pid=69171) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2639}} [ 673.908543] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-22T21:48:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-22T21:48:14Z,direct_url=,disk_format='vmdk',id=8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1f8548ca1c9947769d22274768a59ff7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-22T21:48:15Z,virtual_size=,visibility=), allow threads: False {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 673.908776] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Flavor limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 673.908926] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Image limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 673.909112] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Flavor pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 673.909599] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Image pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 673.909599] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 673.909599] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 673.910909] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 673.911304] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Got 1 possible topologies {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 673.911547] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 673.911717] env[69171]: DEBUG nova.virt.hardware [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 673.912628] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7746306-2f60-42fd-a095-5c33f5772c57 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 673.928176] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d50db0-419c-4155-986f-5b6177fc927a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.099393] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 674.099900] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 674.099900] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 674.099900] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 674.102112] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 674.102112] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 674.102112] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Didn't find any instances for network info cache update. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10015}} [ 674.102112] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.102112] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.102112] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.102737] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.102737] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.102737] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.102737] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=69171) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10548}} [ 674.102737] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager.update_available_resource {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 674.124195] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.306s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.124637] env[69171]: DEBUG nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Start building networks asynchronously for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2830}} [ 674.128445] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.181s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 674.129430] env[69171]: INFO nova.compute.claims [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 674.146711] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235872, 'name': CreateVM_Task, 'duration_secs': 0.361685} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.146711] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Created VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 674.146711] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.146711] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 674.147025] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 674.147472] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-67715b10-b730-49c7-97f0-63f2eb031504 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.155014] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 674.155014] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b23f49-4ba4-f275-501d-b6f6246af6a5" [ 674.155014] env[69171]: _type = "Task" [ 674.155014] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.167433] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b23f49-4ba4-f275-501d-b6f6246af6a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.230416] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235873, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.029223} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.230703] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Deleted the datastore file {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 674.230961] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Moving file from [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 to [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85. {{(pid=69171) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 674.231217] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-ee2aad2b-4163-49fe-bff7-f850dba35989 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.243538] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 674.243538] env[69171]: value = "task-4235874" [ 674.243538] env[69171]: _type = "Task" [ 674.243538] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.255537] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235874, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.608074] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 674.642909] env[69171]: DEBUG nova.compute.utils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Using /dev/sd instead of None {{(pid=69171) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 674.642909] env[69171]: DEBUG nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Allocating IP information in the background. {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1982}} [ 674.642909] env[69171]: DEBUG nova.network.neutron [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] allocate_for_instance() {{(pid=69171) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1156}} [ 674.671232] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.671526] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Processing image 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 674.671788] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 674.761989] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235874, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.032426} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 674.762423] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] File moved {{(pid=69171) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 674.762622] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Cleaning up location [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 674.762780] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Deleting the datastore file [datastore1] vmware_temp/94c494a4-ee92-4a55-83f5-f60a6a1f2d9e {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 674.763066] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-84ffec86-9fa6-4847-9c3b-b4c4b50a7069 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.774699] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 674.774699] env[69171]: value = "task-4235875" [ 674.774699] env[69171]: _type = "Task" [ 674.774699] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 674.789803] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235875, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 674.834482] env[69171]: DEBUG nova.policy [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e0fb9170a8b44c9b843609c3ecbb28da', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b52b9cc0cdf4738b834ecae5a18092f', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=69171) authorize /opt/stack/nova/nova/policy.py:201}} [ 675.133017] env[69171]: DEBUG nova.compute.manager [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Received event network-changed-aad071eb-62cd-4e3d-b8fc-62ea5e458141 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 675.133323] env[69171]: DEBUG nova.compute.manager [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Refreshing instance network info cache due to event network-changed-aad071eb-62cd-4e3d-b8fc-62ea5e458141. {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11137}} [ 675.133471] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Acquiring lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.133584] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Acquired lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.133742] env[69171]: DEBUG nova.network.neutron [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Refreshing network info cache for port aad071eb-62cd-4e3d-b8fc-62ea5e458141 {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.150125] env[69171]: DEBUG nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Start building block device mappings for instance. {{(pid=69171) _build_resources /opt/stack/nova/nova/compute/manager.py:2865}} [ 675.296120] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235875, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.039815} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.299230] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Deleted the datastore file {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 675.300444] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc3c32b9-3d84-4835-b488-425d62b03bad {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.308399] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 675.308399] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52958671-feaa-2788-b9bb-aaa05eb77142" [ 675.308399] env[69171]: _type = "Task" [ 675.308399] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.318704] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52958671-feaa-2788-b9bb-aaa05eb77142, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.329862] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "d8590a5d-0b74-46da-93e4-01cb4d37c7d1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.331761] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "d8590a5d-0b74-46da-93e4-01cb4d37c7d1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.385777] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "5d998742-aa79-4b5b-8448-16ae18fa7b35" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.385777] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "5d998742-aa79-4b5b-8448-16ae18fa7b35" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.445479] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "e00336c6-bcf5-439e-8838-f324423f800d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.445718] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "e00336c6-bcf5-439e-8838-f324423f800d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.602928] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1eb39e61-a552-4656-87a2-8b06d9782dad {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.613354] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d51bc59-4a75-4b1d-82eb-1bee3064a4b7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.666211] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99685c7d-b9de-40fb-b158-c271846a03fe {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.679923] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac64738-0426-439c-b217-b07c3c740746 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.701491] env[69171]: DEBUG nova.compute.provider_tree [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 675.705652] env[69171]: DEBUG nova.compute.manager [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Received event network-changed-8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 675.705652] env[69171]: DEBUG nova.compute.manager [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Refreshing instance network info cache due to event network-changed-8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7. {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11137}} [ 675.706051] env[69171]: DEBUG oslo_concurrency.lockutils [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] Acquiring lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.706226] env[69171]: DEBUG oslo_concurrency.lockutils [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] Acquired lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.706414] env[69171]: DEBUG nova.network.neutron [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Refreshing network info cache for port 8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 675.798069] env[69171]: DEBUG nova.network.neutron [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Successfully updated port: d8a791ab-d322-465f-80d1-73b383b7205b {{(pid=69171) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 675.805051] env[69171]: DEBUG nova.network.neutron [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Successfully created port: 536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 {{(pid=69171) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 675.834350] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52958671-feaa-2788-b9bb-aaa05eb77142, 'name': SearchDatastore_Task, 'duration_secs': 0.013193} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 675.834856] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 675.835291] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] f5a71cca-1e32-4c3c-9bb8-83e634a864f7/f5a71cca-1e32-4c3c-9bb8-83e634a864f7.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 675.836170] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.836482] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 675.837556] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a8ca1739-6e02-4b10-97e0-6062a1e37c2c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.843716] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-97c1e21c-595a-4531-80e4-5f95adb0625c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.852404] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 675.863469] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 675.863469] env[69171]: value = "task-4235876" [ 675.863469] env[69171]: _type = "Task" [ 675.863469] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.866586] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 675.866921] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=69171) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 675.873835] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e19a2075-9147-4c62-b9d1-1ac42172ae79 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.883217] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235876, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 675.884345] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 675.884345] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]5258f89d-0f48-8e0d-4f7c-ceb6712572c9" [ 675.884345] env[69171]: _type = "Task" [ 675.884345] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 675.892882] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]5258f89d-0f48-8e0d-4f7c-ceb6712572c9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.177490] env[69171]: DEBUG nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Start spawning the instance on the hypervisor. {{(pid=69171) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2639}} [ 676.235491] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Getting desirable topologies for flavor Flavor(created_at=2025-05-22T21:48:29Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2025-05-22T21:48:14Z,direct_url=,disk_format='vmdk',id=8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='1f8548ca1c9947769d22274768a59ff7',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2025-05-22T21:48:15Z,virtual_size=,visibility=), allow threads: False {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 676.235741] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Flavor limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 676.235947] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Image limits 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 676.236180] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Flavor pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 676.239014] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Image pref 0:0:0 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 676.239185] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=69171) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 676.239409] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 676.239562] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 676.239729] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Got 1 possible topologies {{(pid=69171) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 676.239905] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 676.240144] env[69171]: DEBUG nova.virt.hardware [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=69171) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 676.241387] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc1cdfb-e2b4-42e6-b69b-00ad840166b9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.244907] env[69171]: ERROR nova.scheduler.client.report [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [req-eb56324b-7585-47d6-9786-ecf6e6fc7244] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-eb56324b-7585-47d6-9786-ecf6e6fc7244"}]} [ 676.245259] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.117s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.248657] env[69171]: ERROR nova.compute.manager [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 676.248657] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Traceback (most recent call last): [ 676.248657] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 676.248657] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] yield [ 676.248657] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 676.248657] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] self.set_inventory_for_provider( [ 676.248657] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 676.248657] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-eb56324b-7585-47d6-9786-ecf6e6fc7244"}]} [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] During handling of the above exception, another exception occurred: [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Traceback (most recent call last): [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] with self.rt.instance_claim(context, instance, node, allocs, [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 676.248948] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] return f(*args, **kwargs) [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] self._update(elevated, cn) [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] self._update_to_placement(context, compute_node, startup) [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] return attempt.get(self._wrap_exception) [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] six.reraise(self.value[0], self.value[1], self.value[2]) [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] raise value [ 676.249240] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] self.reportclient.update_from_provider_tree( [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] with catch_all(pd.uuid): [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] self.gen.throw(typ, value, traceback) [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] raise exception.ResourceProviderSyncFailed() [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 676.249579] env[69171]: ERROR nova.compute.manager [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] [ 676.250455] env[69171]: DEBUG nova.compute.utils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 676.250587] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.375s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.252939] env[69171]: INFO nova.compute.claims [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 676.263366] env[69171]: DEBUG nova.compute.manager [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Build of instance f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 676.263812] env[69171]: DEBUG nova.compute.manager [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 676.264089] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Acquiring lock "refresh_cache-f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.264499] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Acquired lock "refresh_cache-f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.264499] env[69171]: DEBUG nova.network.neutron [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.269738] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fe50ee7-6486-493f-a586-53b89d87c64e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.299308] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "refresh_cache-4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.299360] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquired lock "refresh_cache-4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.299498] env[69171]: DEBUG nova.network.neutron [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 676.384777] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235876, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.395809] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.407754] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]5258f89d-0f48-8e0d-4f7c-ceb6712572c9, 'name': SearchDatastore_Task, 'duration_secs': 0.019475} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.407754] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-80c6446c-6388-4f5a-886b-93f406d8919d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.416577] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 676.416577] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]522e2941-df40-26d6-5660-72909419ef78" [ 676.416577] env[69171]: _type = "Task" [ 676.416577] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.429429] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]522e2941-df40-26d6-5660-72909419ef78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.481522] env[69171]: DEBUG nova.network.neutron [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Updated VIF entry in instance network info cache for port aad071eb-62cd-4e3d-b8fc-62ea5e458141. {{(pid=69171) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 676.481903] env[69171]: DEBUG nova.network.neutron [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Updating instance_info_cache with network_info: [{"id": "aad071eb-62cd-4e3d-b8fc-62ea5e458141", "address": "fa:16:3e:d7:75:e3", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.60", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapaad071eb-62", "ovs_interfaceid": "aad071eb-62cd-4e3d-b8fc-62ea5e458141", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.652979] env[69171]: DEBUG nova.network.neutron [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updated VIF entry in instance network info cache for port 8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7. {{(pid=69171) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 676.653779] env[69171]: DEBUG nova.network.neutron [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updating instance_info_cache with network_info: [{"id": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "address": "fa:16:3e:83:7a:e9", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cd0dd84-ce", "ovs_interfaceid": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.803499] env[69171]: DEBUG nova.network.neutron [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.859696] env[69171]: DEBUG nova.network.neutron [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 676.878555] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235876, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.800129} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.878913] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] f5a71cca-1e32-4c3c-9bb8-83e634a864f7/f5a71cca-1e32-4c3c-9bb8-83e634a864f7.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 676.879177] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Extending root virtual disk to 1048576 {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 676.880210] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0a9427b2-f6c4-48e2-9d4e-6b9d71b03d0f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.888658] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 676.888658] env[69171]: value = "task-4235877" [ 676.888658] env[69171]: _type = "Task" [ 676.888658] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.902906] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235877, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.930209] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]522e2941-df40-26d6-5660-72909419ef78, 'name': SearchDatastore_Task, 'duration_secs': 0.057229} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.930526] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.930795] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] f8b20421-15d3-46f3-a504-172d044ff4d3/f8b20421-15d3-46f3-a504-172d044ff4d3.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 676.931116] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.931306] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 676.931547] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1c61574d-ea4f-4c24-9acd-f4d0da855e7c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.933997] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-16fd200e-e46d-4558-b06c-e728363c4c5c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.937715] env[69171]: DEBUG nova.network.neutron [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.944311] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 676.944311] env[69171]: value = "task-4235878" [ 676.944311] env[69171]: _type = "Task" [ 676.944311] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.944311] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 676.944311] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=69171) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 676.948550] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5bab83d5-98aa-4bf0-a110-73d9967b9818 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.960119] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235878, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 676.960456] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 676.960456] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52f3fbdb-7864-78d8-2158-bf1388676eef" [ 676.960456] env[69171]: _type = "Task" [ 676.960456] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.972056] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52f3fbdb-7864-78d8-2158-bf1388676eef, 'name': SearchDatastore_Task, 'duration_secs': 0.010875} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 676.973317] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aba99ca-7550-4596-a5d2-72b9e19b76e4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.979881] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 676.979881] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52009070-50b1-0f18-1b8a-738c1c8e7902" [ 676.979881] env[69171]: _type = "Task" [ 676.979881] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 676.985151] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Releasing lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.985151] env[69171]: DEBUG nova.compute.manager [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Received event network-vif-plugged-547da9e0-630e-4393-95ef-f0b0086a5b94 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 676.985151] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Acquiring lock "8e6a9d73-6705-484c-82e0-70c6a6860211-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 676.985151] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 676.985151] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.985495] env[69171]: DEBUG nova.compute.manager [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] No waiting events found dispatching network-vif-plugged-547da9e0-630e-4393-95ef-f0b0086a5b94 {{(pid=69171) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 676.985602] env[69171]: WARNING nova.compute.manager [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Received unexpected event network-vif-plugged-547da9e0-630e-4393-95ef-f0b0086a5b94 for instance with vm_state building and task_state spawning. [ 676.985783] env[69171]: DEBUG nova.compute.manager [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Received event network-changed-547da9e0-630e-4393-95ef-f0b0086a5b94 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 676.985893] env[69171]: DEBUG nova.compute.manager [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Refreshing instance network info cache due to event network-changed-547da9e0-630e-4393-95ef-f0b0086a5b94. {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11137}} [ 676.986279] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Acquiring lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 676.986279] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Acquired lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 676.986423] env[69171]: DEBUG nova.network.neutron [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Refreshing network info cache for port 547da9e0-630e-4393-95ef-f0b0086a5b94 {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 676.997852] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52009070-50b1-0f18-1b8a-738c1c8e7902, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.158455] env[69171]: DEBUG oslo_concurrency.lockutils [req-caa421d3-904d-4fd8-b728-80a417414f4b req-4b8bbee6-b1b0-47d7-a0bf-66395fd73027 service nova] Releasing lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.295620] env[69171]: DEBUG nova.scheduler.client.report [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 677.325602] env[69171]: DEBUG nova.scheduler.client.report [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 677.325602] env[69171]: DEBUG nova.compute.provider_tree [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 677.342325] env[69171]: DEBUG nova.scheduler.client.report [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 677.387697] env[69171]: DEBUG nova.scheduler.client.report [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 677.419261] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235877, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.089153} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.419261] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Extended root virtual disk {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 677.421111] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fb5de22-0123-41c5-9e0d-16970724f4eb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.449851] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Releasing lock "refresh_cache-f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.450782] env[69171]: DEBUG nova.compute.manager [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 677.450782] env[69171]: DEBUG nova.compute.manager [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 677.451039] env[69171]: DEBUG nova.network.neutron [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 677.473048] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Reconfiguring VM instance instance-00000002 to attach disk [datastore1] f5a71cca-1e32-4c3c-9bb8-83e634a864f7/f5a71cca-1e32-4c3c-9bb8-83e634a864f7.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 677.483766] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d216d113-ac38-416b-a77c-48de3bcb6fd7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.512109] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235878, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.564683} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.515841] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] f8b20421-15d3-46f3-a504-172d044ff4d3/f8b20421-15d3-46f3-a504-172d044ff4d3.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 677.516200] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Extending root virtual disk to 1048576 {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 677.516538] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 677.516538] env[69171]: value = "task-4235879" [ 677.516538] env[69171]: _type = "Task" [ 677.516538] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.518413] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-59ef0f10-cb78-4114-b097-d07273d278c0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.528138] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52009070-50b1-0f18-1b8a-738c1c8e7902, 'name': SearchDatastore_Task, 'duration_secs': 0.010753} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.528822] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 677.528948] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] 6a29d495-c938-41ad-97b6-1331b6eff589/6a29d495-c938-41ad-97b6-1331b6eff589.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 677.529236] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.529412] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 677.529647] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc639dc8-2d4c-41dc-b962-330ebf687bbf {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.536052] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e09f5e9f-7527-48cb-a793-f4c20309e4a9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.537700] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235879, 'name': ReconfigVM_Task} progress is 14%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.538014] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 677.538014] env[69171]: value = "task-4235880" [ 677.538014] env[69171]: _type = "Task" [ 677.538014] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.547546] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 677.547546] env[69171]: value = "task-4235881" [ 677.547546] env[69171]: _type = "Task" [ 677.547546] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.561429] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 677.561657] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=69171) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 677.562495] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235880, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.566312] env[69171]: DEBUG nova.network.neutron [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 677.568789] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2df14b76-afcc-4bd9-a71a-0d8656220db6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.577534] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235881, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.584516] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 677.584516] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]525d87ae-3b63-4985-9df6-76150cb616cf" [ 677.584516] env[69171]: _type = "Task" [ 677.584516] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.611876] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]525d87ae-3b63-4985-9df6-76150cb616cf, 'name': SearchDatastore_Task, 'duration_secs': 0.010955} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 677.611876] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-732451f3-3073-4d11-a913-541da4936972 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.615738] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 677.615738] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b1aec1-7b3f-0606-208a-236fd48e505d" [ 677.615738] env[69171]: _type = "Task" [ 677.615738] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 677.627449] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b1aec1-7b3f-0606-208a-236fd48e505d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 677.742935] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-139e1b7f-8b66-42b3-9c9f-01670263f366 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.751981] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23a6f9ab-fa13-4cbf-ac57-5110f16dca73 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.786747] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d2f5963-56a6-4af9-9d07-79dfc33987ea {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.796890] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-144b64a6-b7f5-4f70-8540-b0fcc5474f91 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.816025] env[69171]: DEBUG nova.compute.provider_tree [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 677.832555] env[69171]: DEBUG nova.network.neutron [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Updating instance_info_cache with network_info: [{"id": "d8a791ab-d322-465f-80d1-73b383b7205b", "address": "fa:16:3e:ba:c7:27", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8a791ab-d3", "ovs_interfaceid": "d8a791ab-d322-465f-80d1-73b383b7205b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.035355] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235879, 'name': ReconfigVM_Task} progress is 14%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.052721] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235880, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075229} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.056438] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Extended root virtual disk {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 678.057023] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1479d03-d9d9-4122-bbcb-d807ae8f046a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.066675] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235881, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.489883} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.078968] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] 6a29d495-c938-41ad-97b6-1331b6eff589/6a29d495-c938-41ad-97b6-1331b6eff589.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 678.079124] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Extending root virtual disk to 1048576 {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 678.079496] env[69171]: DEBUG nova.network.neutron [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.092472] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Reconfiguring VM instance instance-00000001 to attach disk [datastore1] f8b20421-15d3-46f3-a504-172d044ff4d3/f8b20421-15d3-46f3-a504-172d044ff4d3.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 678.093829] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed54bb43-162b-4e14-8d9f-af3055722197 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.096224] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e3d0cf1-1188-45c5-817d-77b80bbff331 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.123204] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 678.123204] env[69171]: value = "task-4235882" [ 678.123204] env[69171]: _type = "Task" [ 678.123204] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.132866] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b1aec1-7b3f-0606-208a-236fd48e505d, 'name': SearchDatastore_Task, 'duration_secs': 0.010502} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.133248] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 678.133248] env[69171]: value = "task-4235883" [ 678.133248] env[69171]: _type = "Task" [ 678.133248] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.135305] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.135305] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] 8e6a9d73-6705-484c-82e0-70c6a6860211/8e6a9d73-6705-484c-82e0-70c6a6860211.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 678.135305] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5795b35d-538b-415e-a8cb-43d3e28b4068 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.146990] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235882, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.157366] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235883, 'name': ReconfigVM_Task} progress is 14%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.157366] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 678.157366] env[69171]: value = "task-4235884" [ 678.157366] env[69171]: _type = "Task" [ 678.157366] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.165867] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235884, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.204497] env[69171]: DEBUG nova.network.neutron [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Successfully updated port: 536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 {{(pid=69171) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 678.291774] env[69171]: DEBUG nova.network.neutron [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Updated VIF entry in instance network info cache for port 547da9e0-630e-4393-95ef-f0b0086a5b94. {{(pid=69171) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 678.292206] env[69171]: DEBUG nova.network.neutron [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Updating instance_info_cache with network_info: [{"id": "547da9e0-630e-4393-95ef-f0b0086a5b94", "address": "fa:16:3e:4f:61:19", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.88", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap547da9e0-63", "ovs_interfaceid": "547da9e0-630e-4393-95ef-f0b0086a5b94", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.336596] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Releasing lock "refresh_cache-4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.337045] env[69171]: DEBUG nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Instance network_info: |[{"id": "d8a791ab-d322-465f-80d1-73b383b7205b", "address": "fa:16:3e:ba:c7:27", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8a791ab-d3", "ovs_interfaceid": "d8a791ab-d322-465f-80d1-73b383b7205b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1997}} [ 678.337523] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ba:c7:27', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd8a791ab-d322-465f-80d1-73b383b7205b', 'vif_model': 'vmxnet3'}] {{(pid=69171) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 678.346792] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Creating folder: Project (f851949057dc4a268959f6c7cf0bd632). Parent ref: group-v838162. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 678.347896] env[69171]: ERROR nova.scheduler.client.report [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [req-b767810f-7207-44ed-9eeb-e20c7c194de2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b767810f-7207-44ed-9eeb-e20c7c194de2"}]} [ 678.348076] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.097s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 678.348711] env[69171]: ERROR nova.compute.manager [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 678.348711] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Traceback (most recent call last): [ 678.348711] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 678.348711] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] yield [ 678.348711] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 678.348711] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] self.set_inventory_for_provider( [ 678.348711] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 678.348711] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b767810f-7207-44ed-9eeb-e20c7c194de2"}]} [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] During handling of the above exception, another exception occurred: [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Traceback (most recent call last): [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] with self.rt.instance_claim(context, instance, node, allocs, [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 678.349120] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] return f(*args, **kwargs) [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] self._update(elevated, cn) [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] self._update_to_placement(context, compute_node, startup) [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] return attempt.get(self._wrap_exception) [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] six.reraise(self.value[0], self.value[1], self.value[2]) [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] raise value [ 678.349558] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] self.reportclient.update_from_provider_tree( [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] with catch_all(pd.uuid): [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] self.gen.throw(typ, value, traceback) [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] raise exception.ResourceProviderSyncFailed() [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 678.350157] env[69171]: ERROR nova.compute.manager [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] [ 678.350726] env[69171]: DEBUG nova.compute.utils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 678.351113] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e750db5f-20a3-41b7-a7b9-6160fe96efbc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.362028] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.309s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.362028] env[69171]: INFO nova.compute.claims [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 678.364856] env[69171]: DEBUG nova.compute.manager [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Build of instance 23e40e0a-45e2-4735-b953-2c165d318c50 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 678.368813] env[69171]: DEBUG nova.compute.manager [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 678.368942] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "refresh_cache-23e40e0a-45e2-4735-b953-2c165d318c50" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.369050] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquired lock "refresh_cache-23e40e0a-45e2-4735-b953-2c165d318c50" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.369219] env[69171]: DEBUG nova.network.neutron [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.383627] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Created folder: Project (f851949057dc4a268959f6c7cf0bd632) in parent group-v838162. [ 678.383759] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Creating folder: Instances. Parent ref: group-v838175. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 678.384752] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e2973359-6843-4944-9aa4-1bef8a51b64c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.397422] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Created folder: Instances in parent group-v838175. [ 678.397706] env[69171]: DEBUG oslo.service.loopingcall [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 678.399362] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Creating VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 678.399362] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ba7ab8de-ba6f-47b3-8e62-b989831a8960 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.421381] env[69171]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 678.421381] env[69171]: value = "task-4235887" [ 678.421381] env[69171]: _type = "Task" [ 678.421381] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.433671] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235887, 'name': CreateVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.534882] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235879, 'name': ReconfigVM_Task, 'duration_secs': 0.800584} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.535342] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Reconfigured VM instance instance-00000002 to attach disk [datastore1] f5a71cca-1e32-4c3c-9bb8-83e634a864f7/f5a71cca-1e32-4c3c-9bb8-83e634a864f7.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 678.536169] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-84cf0f1a-0677-40d1-a62f-0634ea296ff8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.544204] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 678.544204] env[69171]: value = "task-4235888" [ 678.544204] env[69171]: _type = "Task" [ 678.544204] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.559497] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235888, 'name': Rename_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.594538] env[69171]: INFO nova.compute.manager [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] [instance: f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9] Took 1.14 seconds to deallocate network for instance. [ 678.638526] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235882, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.084103} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.647126] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Extended root virtual disk {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 678.647126] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91859380-deed-491c-839e-3d0a09ccbf38 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.655718] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235883, 'name': ReconfigVM_Task} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.664361] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Reconfigured VM instance instance-00000001 to attach disk [datastore1] f8b20421-15d3-46f3-a504-172d044ff4d3/f8b20421-15d3-46f3-a504-172d044ff4d3.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 678.675491] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Reconfiguring VM instance instance-00000004 to attach disk [datastore1] 6a29d495-c938-41ad-97b6-1331b6eff589/6a29d495-c938-41ad-97b6-1331b6eff589.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 678.675858] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-60bfa582-d546-4e4b-8ec0-8b3955fabf2e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.684554] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cd99707f-a844-425f-88aa-c52c3566a186 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.704499] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235884, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.499137} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.708304] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] 8e6a9d73-6705-484c-82e0-70c6a6860211/8e6a9d73-6705-484c-82e0-70c6a6860211.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 678.708304] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Extending root virtual disk to 1048576 {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 678.708304] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 678.708304] env[69171]: value = "task-4235890" [ 678.708304] env[69171]: _type = "Task" [ 678.708304] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.708304] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 678.708304] env[69171]: value = "task-4235889" [ 678.708304] env[69171]: _type = "Task" [ 678.708304] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.708994] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2bb4406a-7a33-4d6c-a29e-c48ee6ca06d5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.710566] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.710718] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquired lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.710872] env[69171]: DEBUG nova.network.neutron [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 678.734565] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235889, 'name': Rename_Task} progress is 14%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.735771] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235890, 'name': ReconfigVM_Task} progress is 14%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.736849] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 678.736849] env[69171]: value = "task-4235891" [ 678.736849] env[69171]: _type = "Task" [ 678.736849] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.750915] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235891, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 678.799189] env[69171]: DEBUG oslo_concurrency.lockutils [req-c1f025b1-7c1f-42f4-8fb7-49c1bbd692a5 req-531e7cbe-e942-4b31-b7f9-67d413df314e service nova] Releasing lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.894040] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Acquiring lock "62413635-6f32-4539-b9ee-6ebe50733cf1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.894040] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Lock "62413635-6f32-4539-b9ee-6ebe50733cf1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 678.933160] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235887, 'name': CreateVM_Task, 'duration_secs': 0.441702} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 678.934215] env[69171]: DEBUG nova.network.neutron [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 678.936108] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Created VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 678.936861] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 678.937060] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.937474] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 678.938051] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8717c4e8-c3a4-4288-9a4c-6da348f01175 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.943840] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 678.943840] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]5227bddd-87bc-5a14-e0af-90f02f8c724f" [ 678.943840] env[69171]: _type = "Task" [ 678.943840] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 678.953188] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]5227bddd-87bc-5a14-e0af-90f02f8c724f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.059039] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235888, 'name': Rename_Task, 'duration_secs': 0.25973} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.059691] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Powering on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 679.060124] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b1e058db-e7fc-423e-99be-1542243e17b4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.069396] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 679.069396] env[69171]: value = "task-4235892" [ 679.069396] env[69171]: _type = "Task" [ 679.069396] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.087517] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.211216] env[69171]: DEBUG nova.network.neutron [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.237442] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235889, 'name': Rename_Task, 'duration_secs': 0.175197} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.237707] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235890, 'name': ReconfigVM_Task, 'duration_secs': 0.402893} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.238043] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Powering on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 679.238657] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Reconfigured VM instance instance-00000004 to attach disk [datastore1] 6a29d495-c938-41ad-97b6-1331b6eff589/6a29d495-c938-41ad-97b6-1331b6eff589.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 679.244541] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e6ca177-caff-45b1-90f4-3893cc34b3b2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.248155] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93dcae26-117c-406c-90c3-2543d0aebc85 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.258472] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235891, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.092757} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.261207] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Extended root virtual disk {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 679.261971] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 679.261971] env[69171]: value = "task-4235893" [ 679.261971] env[69171]: _type = "Task" [ 679.261971] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.262253] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 679.262253] env[69171]: value = "task-4235894" [ 679.262253] env[69171]: _type = "Task" [ 679.262253] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.263082] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d118973-38c2-4662-9c1b-706ab56f58d7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.279452] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235894, 'name': Rename_Task} progress is 10%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.292012] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235893, 'name': PowerOnVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.300784] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Reconfiguring VM instance instance-00000003 to attach disk [datastore1] 8e6a9d73-6705-484c-82e0-70c6a6860211/8e6a9d73-6705-484c-82e0-70c6a6860211.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 679.301220] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6529610b-fa93-4296-84b2-40c4ef3106ee {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.323152] env[69171]: DEBUG nova.network.neutron [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.326968] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 679.326968] env[69171]: value = "task-4235895" [ 679.326968] env[69171]: _type = "Task" [ 679.326968] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.338937] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235895, 'name': ReconfigVM_Task} progress is 6%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.416165] env[69171]: DEBUG nova.scheduler.client.report [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 679.441843] env[69171]: DEBUG nova.scheduler.client.report [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 679.442164] env[69171]: DEBUG nova.compute.provider_tree [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 679.464705] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]5227bddd-87bc-5a14-e0af-90f02f8c724f, 'name': SearchDatastore_Task, 'duration_secs': 0.013701} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.466050] env[69171]: DEBUG nova.scheduler.client.report [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 679.468778] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.468778] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Processing image 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 679.468778] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.468778] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.468940] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 679.470609] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e6b0f09f-011a-48c9-ac19-12243b912552 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.485917] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 679.487412] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=69171) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 679.490517] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49e2ff2e-773a-41d1-934c-be65b78b4c12 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.496958] env[69171]: DEBUG nova.scheduler.client.report [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 679.501686] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 679.501686] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]5254472a-9394-f89e-f1b0-b2c555093425" [ 679.501686] env[69171]: _type = "Task" [ 679.501686] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.513693] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]5254472a-9394-f89e-f1b0-b2c555093425, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.583501] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235892, 'name': PowerOnVM_Task} progress is 89%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.642896] env[69171]: INFO nova.scheduler.client.report [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Deleted allocations for instance f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9 [ 679.715352] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Releasing lock "refresh_cache-23e40e0a-45e2-4735-b953-2c165d318c50" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.715618] env[69171]: DEBUG nova.compute.manager [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 679.715851] env[69171]: DEBUG nova.compute.manager [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 679.716046] env[69171]: DEBUG nova.network.neutron [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 679.758613] env[69171]: DEBUG nova.network.neutron [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 679.784974] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235894, 'name': Rename_Task, 'duration_secs': 0.197931} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 679.789500] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Powering on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 679.790496] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235893, 'name': PowerOnVM_Task} progress is 89%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.791031] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4654e963-d171-441a-b4d8-5fdd8d448945 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.801240] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9db99b1-fb02-4972-9f13-d5e5054391d8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.806992] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 679.806992] env[69171]: value = "task-4235896" [ 679.806992] env[69171]: _type = "Task" [ 679.806992] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 679.814793] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64588611-aa8a-49ea-9d2f-1bac6975c584 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.855191] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.859884] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b6e683-c7ed-4e68-ab11-ffc8d2784014 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.872574] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffea4606-96bc-4d96-84e6-b2109225c505 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.877099] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235895, 'name': ReconfigVM_Task} progress is 99%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 679.892057] env[69171]: DEBUG nova.compute.provider_tree [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 680.018557] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]5254472a-9394-f89e-f1b0-b2c555093425, 'name': SearchDatastore_Task, 'duration_secs': 0.016619} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.019770] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3aa8d01e-76fd-4808-a04b-1ecea5c489de {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.027573] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 680.027573] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]526d432f-3e92-3bcd-f618-2280c1337a2d" [ 680.027573] env[69171]: _type = "Task" [ 680.027573] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.038665] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]526d432f-3e92-3bcd-f618-2280c1337a2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.055975] env[69171]: DEBUG nova.network.neutron [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Updating instance_info_cache with network_info: [{"id": "536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0", "address": "fa:16:3e:10:14:88", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.82", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap536e9ab7-6a", "ovs_interfaceid": "536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.085358] env[69171]: DEBUG oslo_vmware.api [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235892, 'name': PowerOnVM_Task, 'duration_secs': 0.563498} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.085358] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Powered on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 680.085358] env[69171]: INFO nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Took 15.90 seconds to spawn the instance on the hypervisor. [ 680.085358] env[69171]: DEBUG nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 680.085358] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f020ced-96ab-48fe-8fd5-5593e0a0100b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.156622] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3d2908e1-9863-4299-90d7-d5b73c612bdc tempest-ServerExternalEventsTest-225465277 tempest-ServerExternalEventsTest-225465277-project-member] Lock "f4ac66bd-c113-4c9f-a0ac-8eadbeba4de9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.254s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.213207] env[69171]: DEBUG nova.compute.manager [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Received event network-vif-plugged-d8a791ab-d322-465f-80d1-73b383b7205b {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 680.213207] env[69171]: DEBUG oslo_concurrency.lockutils [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] Acquiring lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.213207] env[69171]: DEBUG oslo_concurrency.lockutils [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.213207] env[69171]: DEBUG oslo_concurrency.lockutils [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.216170] env[69171]: DEBUG nova.compute.manager [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] No waiting events found dispatching network-vif-plugged-d8a791ab-d322-465f-80d1-73b383b7205b {{(pid=69171) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 680.216170] env[69171]: WARNING nova.compute.manager [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Received unexpected event network-vif-plugged-d8a791ab-d322-465f-80d1-73b383b7205b for instance with vm_state building and task_state spawning. [ 680.216170] env[69171]: DEBUG nova.compute.manager [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Received event network-changed-d8a791ab-d322-465f-80d1-73b383b7205b {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 680.216170] env[69171]: DEBUG nova.compute.manager [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Refreshing instance network info cache due to event network-changed-d8a791ab-d322-465f-80d1-73b383b7205b. {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11137}} [ 680.216170] env[69171]: DEBUG oslo_concurrency.lockutils [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] Acquiring lock "refresh_cache-4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.216429] env[69171]: DEBUG oslo_concurrency.lockutils [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] Acquired lock "refresh_cache-4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.216801] env[69171]: DEBUG nova.network.neutron [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Refreshing network info cache for port d8a791ab-d322-465f-80d1-73b383b7205b {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 680.266373] env[69171]: DEBUG nova.network.neutron [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.281348] env[69171]: DEBUG oslo_vmware.api [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235893, 'name': PowerOnVM_Task, 'duration_secs': 0.566716} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.281731] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Powered on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 680.281940] env[69171]: INFO nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Took 18.31 seconds to spawn the instance on the hypervisor. [ 680.282140] env[69171]: DEBUG nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 680.284126] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93180ed1-64d9-45e2-b5da-3f4e52f73cb1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.324023] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235896, 'name': PowerOnVM_Task} progress is 64%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.369498] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235895, 'name': ReconfigVM_Task, 'duration_secs': 0.586831} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.369809] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Reconfigured VM instance instance-00000003 to attach disk [datastore1] 8e6a9d73-6705-484c-82e0-70c6a6860211/8e6a9d73-6705-484c-82e0-70c6a6860211.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 680.371891] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-13838d12-4043-4f7c-a9f1-aafc57ac7005 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.379162] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 680.379162] env[69171]: value = "task-4235897" [ 680.379162] env[69171]: _type = "Task" [ 680.379162] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.391166] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235897, 'name': Rename_Task} progress is 5%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.426273] env[69171]: ERROR nova.scheduler.client.report [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [req-d2a08620-b6a1-4b69-8025-19b92e76699b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d2a08620-b6a1-4b69-8025-19b92e76699b"}]} [ 680.426609] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.067s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.427233] env[69171]: ERROR nova.compute.manager [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 680.427233] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Traceback (most recent call last): [ 680.427233] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 680.427233] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] yield [ 680.427233] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 680.427233] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] self.set_inventory_for_provider( [ 680.427233] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 680.427233] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d2a08620-b6a1-4b69-8025-19b92e76699b"}]} [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] During handling of the above exception, another exception occurred: [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Traceback (most recent call last): [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] with self.rt.instance_claim(context, instance, node, allocs, [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 680.427532] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] return f(*args, **kwargs) [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] self._update(elevated, cn) [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] self._update_to_placement(context, compute_node, startup) [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] return attempt.get(self._wrap_exception) [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] six.reraise(self.value[0], self.value[1], self.value[2]) [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] raise value [ 680.427803] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] self.reportclient.update_from_provider_tree( [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] with catch_all(pd.uuid): [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] self.gen.throw(typ, value, traceback) [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] raise exception.ResourceProviderSyncFailed() [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 680.428180] env[69171]: ERROR nova.compute.manager [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] [ 680.428448] env[69171]: DEBUG nova.compute.utils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 680.433162] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 5.821s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.434232] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.004s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 680.434232] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=69171) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 680.434232] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.038s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.435669] env[69171]: INFO nova.compute.claims [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.439722] env[69171]: DEBUG nova.compute.manager [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Build of instance 836c5a40-0325-49d5-90d2-bd86cb3748ab was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 680.441452] env[69171]: DEBUG nova.compute.manager [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 680.441674] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Acquiring lock "refresh_cache-836c5a40-0325-49d5-90d2-bd86cb3748ab" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 680.441955] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Acquired lock "refresh_cache-836c5a40-0325-49d5-90d2-bd86cb3748ab" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 680.442032] env[69171]: DEBUG nova.network.neutron [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 680.443910] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-403189e1-987b-4dc6-9157-c73c0871a84f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.459552] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f6ea0d-f24d-44d0-b64c-5edc8214a7b9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.476728] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f094aa45-cc35-4977-b814-eeb6a188659c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.490917] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a3362d-d36a-45cb-821a-b53982d4a563 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.540279] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180700MB free_disk=0GB free_vcpus=48 pci_devices=None {{(pid=69171) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 680.540279] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.550239] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]526d432f-3e92-3bcd-f618-2280c1337a2d, 'name': SearchDatastore_Task, 'duration_secs': 0.011259} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.550515] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.550762] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03/4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 680.551037] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-868d2986-138f-4a61-a321-b52ca655ddcd {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.558626] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Releasing lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.558909] env[69171]: DEBUG nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance network_info: |[{"id": "536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0", "address": "fa:16:3e:10:14:88", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.82", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap536e9ab7-6a", "ovs_interfaceid": "536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=69171) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1997}} [ 680.559266] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 680.559266] env[69171]: value = "task-4235898" [ 680.559266] env[69171]: _type = "Task" [ 680.559266] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.559583] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:10:14:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f17856cf-7248-414b-bde6-8c90cfb4c593', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0', 'vif_model': 'vmxnet3'}] {{(pid=69171) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 680.567185] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Creating folder: Project (7b52b9cc0cdf4738b834ecae5a18092f). Parent ref: group-v838162. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 680.567880] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a740eaa9-e122-4d35-abea-d6b46b645f05 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.578851] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235898, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.582853] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Created folder: Project (7b52b9cc0cdf4738b834ecae5a18092f) in parent group-v838162. [ 680.583791] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Creating folder: Instances. Parent ref: group-v838178. {{(pid=69171) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1589}} [ 680.583791] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-33371ac3-75ff-4a28-9aab-e192a5ca6006 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.594406] env[69171]: INFO nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Created folder: Instances in parent group-v838178. [ 680.594517] env[69171]: DEBUG oslo.service.loopingcall [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 680.594723] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Creating VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1327}} [ 680.595038] env[69171]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5dc262d6-6d43-41ff-80bf-2fdbb122cad3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.620592] env[69171]: INFO nova.compute.manager [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Took 22.90 seconds to build instance. [ 680.625413] env[69171]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 680.625413] env[69171]: value = "task-4235901" [ 680.625413] env[69171]: _type = "Task" [ 680.625413] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.636169] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235901, 'name': CreateVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 680.661556] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 680.769417] env[69171]: INFO nova.compute.manager [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 23e40e0a-45e2-4735-b953-2c165d318c50] Took 1.05 seconds to deallocate network for instance. [ 680.806069] env[69171]: INFO nova.compute.manager [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Took 23.09 seconds to build instance. [ 680.826266] env[69171]: DEBUG oslo_vmware.api [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235896, 'name': PowerOnVM_Task, 'duration_secs': 0.974229} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.827487] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Powered on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 680.827487] env[69171]: INFO nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Took 9.31 seconds to spawn the instance on the hypervisor. [ 680.827487] env[69171]: DEBUG nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 680.827924] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a67656d-c4dd-464f-b353-cc0e8d0b0f6d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.894695] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235897, 'name': Rename_Task, 'duration_secs': 0.178778} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 680.894850] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Powering on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 680.896080] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-242a0be8-da13-4717-8b70-eae7d3e5a1b1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.907116] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 680.907116] env[69171]: value = "task-4235902" [ 680.907116] env[69171]: _type = "Task" [ 680.907116] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 680.916983] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235902, 'name': PowerOnVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.017207] env[69171]: DEBUG nova.network.neutron [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.084660] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235898, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.124816] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f5755304-a36f-4ece-834e-ebbce3b0d27f tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.412s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.140830] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235901, 'name': CreateVM_Task} progress is 25%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.192922] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 681.227205] env[69171]: DEBUG nova.network.neutron [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.316825] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d666a11-8488-41cc-9a37-c50cc5dcbc8f tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.609s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.356950] env[69171]: INFO nova.compute.manager [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Took 14.17 seconds to build instance. [ 681.422527] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235902, 'name': PowerOnVM_Task} progress is 66%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.482662] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 681.502378] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 681.502537] env[69171]: DEBUG nova.compute.provider_tree [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa generation from 17 to 18 during operation: update_inventory {{(pid=69171) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 681.502692] env[69171]: DEBUG nova.compute.provider_tree [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 681.515825] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 681.548168] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 681.581208] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235898, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.707819} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.584184] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03/4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 681.584302] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Extending root virtual disk to 1048576 {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 681.584789] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3bb6d00d-16d1-46df-808b-6a7202d4085a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.595941] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 681.595941] env[69171]: value = "task-4235903" [ 681.595941] env[69171]: _type = "Task" [ 681.595941] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.611890] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235903, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.612917] env[69171]: DEBUG nova.network.neutron [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Updated VIF entry in instance network info cache for port d8a791ab-d322-465f-80d1-73b383b7205b. {{(pid=69171) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 681.613337] env[69171]: DEBUG nova.network.neutron [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Updating instance_info_cache with network_info: [{"id": "d8a791ab-d322-465f-80d1-73b383b7205b", "address": "fa:16:3e:ba:c7:27", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.105", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd8a791ab-d3", "ovs_interfaceid": "d8a791ab-d322-465f-80d1-73b383b7205b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.633573] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 681.640675] env[69171]: DEBUG oslo_vmware.api [-] Task: {'id': task-4235901, 'name': CreateVM_Task, 'duration_secs': 0.732066} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 681.641548] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Created VM on the ESX host {{(pid=69171) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1349}} [ 681.641789] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 681.641945] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 681.643126] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 681.643126] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7f438e52-6a90-4ae1-9e1d-acbde166bd11 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.650706] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 681.650706] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52dd20e4-2e5e-19ab-422e-beb0c8347f14" [ 681.650706] env[69171]: _type = "Task" [ 681.650706] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 681.662408] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52dd20e4-2e5e-19ab-422e-beb0c8347f14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 681.731665] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Releasing lock "refresh_cache-836c5a40-0325-49d5-90d2-bd86cb3748ab" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 681.731872] env[69171]: DEBUG nova.compute.manager [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 681.732089] env[69171]: DEBUG nova.compute.manager [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 681.732306] env[69171]: DEBUG nova.network.neutron [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 681.765791] env[69171]: DEBUG nova.network.neutron [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 681.809686] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a1728e7-45cc-4fba-b090-22358c577a4d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.818580] env[69171]: INFO nova.scheduler.client.report [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Deleted allocations for instance 23e40e0a-45e2-4735-b953-2c165d318c50 [ 681.825891] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35d4640-e086-43b5-a77e-7f6c472b1d75 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.829501] env[69171]: DEBUG nova.compute.manager [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 681.868870] env[69171]: DEBUG oslo_concurrency.lockutils [None req-d904ba71-10c6-4b3f-b2a2-c58c578952d8 tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "6a29d495-c938-41ad-97b6-1331b6eff589" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.708s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.869982] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32b775b7-09df-4ba3-a116-a54192fd0b54 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.880553] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8231fed-46bc-43b1-a7c1-b628d7569744 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.896142] env[69171]: DEBUG nova.compute.provider_tree [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 681.921815] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235902, 'name': PowerOnVM_Task} progress is 100%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.064093] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Acquiring lock "eb6e8ba7-7ebd-48ec-b536-7e0468011057" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.064658] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Lock "eb6e8ba7-7ebd-48ec-b536-7e0468011057" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.112245] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235903, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074267} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.112703] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Extended root virtual disk {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 682.113496] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de691710-a4f6-4b29-8d7d-da55e1003b5f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.119628] env[69171]: DEBUG oslo_concurrency.lockutils [req-16a052dd-c538-4697-9bf4-f68762661efa req-f03e4fe2-b03d-460a-8943-0b186c0662d2 service nova] Releasing lock "refresh_cache-4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.144506] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Reconfiguring VM instance instance-00000005 to attach disk [datastore1] 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03/4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 682.147698] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-734e25d3-1776-48ca-842a-74b53360699b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.189542] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52dd20e4-2e5e-19ab-422e-beb0c8347f14, 'name': SearchDatastore_Task, 'duration_secs': 0.013246} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.191300] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 682.191507] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Processing image 8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85 {{(pid=69171) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 682.191769] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.191913] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquired lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.192105] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 682.192479] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 682.192479] env[69171]: value = "task-4235904" [ 682.192479] env[69171]: _type = "Task" [ 682.192479] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.192698] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a14844bc-ac1a-4b9f-b55e-56228158d1b8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.208567] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235904, 'name': ReconfigVM_Task} progress is 14%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.211039] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.213282] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=69171) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 682.213515] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=69171) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1851}} [ 682.214619] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9c87a666-e9aa-4687-ac6c-bf49bdd56cc5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.223210] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 682.223210] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b2dec6-5e55-3740-04c3-a64f8083aa78" [ 682.223210] env[69171]: _type = "Task" [ 682.223210] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.233138] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b2dec6-5e55-3740-04c3-a64f8083aa78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.272212] env[69171]: DEBUG nova.network.neutron [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.337808] env[69171]: DEBUG oslo_concurrency.lockutils [None req-aaf260d0-cc63-422d-8a27-60af29733529 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "23e40e0a-45e2-4735-b953-2c165d318c50" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.532s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.356802] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.418449] env[69171]: DEBUG oslo_vmware.api [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235902, 'name': PowerOnVM_Task, 'duration_secs': 1.025154} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.419737] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Powered on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 682.419737] env[69171]: INFO nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Took 13.73 seconds to spawn the instance on the hypervisor. [ 682.419737] env[69171]: DEBUG nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 682.419974] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-548c69e9-4bb3-4352-bd74-61f46b61fa7f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.437246] env[69171]: ERROR nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [req-01639872-c129-4eea-bf79-7d4cf991eb0b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-01639872-c129-4eea-bf79-7d4cf991eb0b"}]} [ 682.437647] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.004s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.438277] env[69171]: ERROR nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 682.438277] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Traceback (most recent call last): [ 682.438277] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 682.438277] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] yield [ 682.438277] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 682.438277] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] self.set_inventory_for_provider( [ 682.438277] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 682.438277] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-01639872-c129-4eea-bf79-7d4cf991eb0b"}]} [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] During handling of the above exception, another exception occurred: [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Traceback (most recent call last): [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] with self.rt.instance_claim(context, instance, node, allocs, [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 682.438568] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] return f(*args, **kwargs) [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] self._update(elevated, cn) [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] self._update_to_placement(context, compute_node, startup) [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] return attempt.get(self._wrap_exception) [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] six.reraise(self.value[0], self.value[1], self.value[2]) [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] raise value [ 682.438848] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] self.reportclient.update_from_provider_tree( [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] with catch_all(pd.uuid): [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] self.gen.throw(typ, value, traceback) [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] raise exception.ResourceProviderSyncFailed() [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 682.439302] env[69171]: ERROR nova.compute.manager [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] [ 682.439596] env[69171]: DEBUG nova.compute.utils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 682.444670] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.901s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.444670] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Build of instance d8590a5d-0b74-46da-93e4-01cb4d37c7d1 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 682.444670] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 682.444670] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "refresh_cache-d8590a5d-0b74-46da-93e4-01cb4d37c7d1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.445113] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquired lock "refresh_cache-d8590a5d-0b74-46da-93e4-01cb4d37c7d1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.445113] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 682.567351] env[69171]: DEBUG nova.compute.manager [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 682.706689] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235904, 'name': ReconfigVM_Task, 'duration_secs': 0.491534} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.706689] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Reconfigured VM instance instance-00000005 to attach disk [datastore1] 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03/4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 682.707594] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-65229067-d461-4b8f-aea1-90b331099507 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.715300] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 682.715300] env[69171]: value = "task-4235905" [ 682.715300] env[69171]: _type = "Task" [ 682.715300] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.725132] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235905, 'name': Rename_Task} progress is 5%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.735132] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52b2dec6-5e55-3740-04c3-a64f8083aa78, 'name': SearchDatastore_Task, 'duration_secs': 0.030211} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 682.736062] env[69171]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fca96cd8-0c1c-4e9f-81e1-9459c37d3f92 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.745397] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 682.745397] env[69171]: value = "session[52231107-d32e-c993-fc9b-e6b37fa56ced]52c73d99-3416-0de9-1daa-2c708e89573a" [ 682.745397] env[69171]: _type = "Task" [ 682.745397] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 682.757192] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52c73d99-3416-0de9-1daa-2c708e89573a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 682.777809] env[69171]: INFO nova.compute.manager [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] [instance: 836c5a40-0325-49d5-90d2-bd86cb3748ab] Took 1.04 seconds to deallocate network for instance. [ 682.962157] env[69171]: INFO nova.compute.manager [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Took 18.55 seconds to build instance. [ 682.987762] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.096860] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.228557] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235905, 'name': Rename_Task, 'duration_secs': 0.292303} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.229619] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 683.230841] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Powering on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 683.231255] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7634673a-5791-40a6-9cf8-322692e49be2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.240026] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 683.240026] env[69171]: value = "task-4235906" [ 683.240026] env[69171]: _type = "Task" [ 683.240026] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.256244] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235906, 'name': PowerOnVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.260573] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': session[52231107-d32e-c993-fc9b-e6b37fa56ced]52c73d99-3416-0de9-1daa-2c708e89573a, 'name': SearchDatastore_Task, 'duration_secs': 0.016876} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 683.260914] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Releasing lock "[datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.261244] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] fa5997fa-817c-42b6-a7fa-b3b85a109b6d/fa5997fa-817c-42b6-a7fa-b3b85a109b6d.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1413}} [ 683.261571] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fbe7b4d7-9dfb-4a52-be92-ce95f58296b2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.273470] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 683.273470] env[69171]: value = "task-4235907" [ 683.273470] env[69171]: _type = "Task" [ 683.273470] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 683.286518] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235907, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.466705] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ce929ab5-2cab-4dd1-a78e-63541ce06cfc tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.068s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.491116] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f8b20421-15d3-46f3-a504-172d044ff4d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.491116] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f5a71cca-1e32-4c3c-9bb8-83e634a864f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.491116] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 8e6a9d73-6705-484c-82e0-70c6a6860211 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.491116] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 6a29d495-c938-41ad-97b6-1331b6eff589 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.491361] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.491361] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance fa5997fa-817c-42b6-a7fa-b3b85a109b6d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.491361] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 836c5a40-0325-49d5-90d2-bd86cb3748ab actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.491361] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance d8590a5d-0b74-46da-93e4-01cb4d37c7d1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 683.720325] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Acquiring lock "dffe1923-6324-415c-9e38-680e2b213dca" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.721818] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Lock "dffe1923-6324-415c-9e38-680e2b213dca" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.736392] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Releasing lock "refresh_cache-d8590a5d-0b74-46da-93e4-01cb4d37c7d1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.736392] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 683.736392] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 683.736392] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 683.754957] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235906, 'name': PowerOnVM_Task} progress is 89%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.771491] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 683.791355] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235907, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 683.830415] env[69171]: INFO nova.scheduler.client.report [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Deleted allocations for instance 836c5a40-0325-49d5-90d2-bd86cb3748ab [ 683.995161] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 5d998742-aa79-4b5b-8448-16ae18fa7b35 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.091347] env[69171]: DEBUG nova.compute.manager [None req-fe3a6945-9e2d-4a94-bc68-64bf651ec741 tempest-ServerDiagnosticsV248Test-883600288 tempest-ServerDiagnosticsV248Test-883600288-project-admin] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 684.095707] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38296f5e-5836-4140-8708-748e0fc94133 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.102859] env[69171]: DEBUG nova.compute.manager [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Received event network-vif-plugged-536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 684.103097] env[69171]: DEBUG oslo_concurrency.lockutils [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] Acquiring lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.103291] env[69171]: DEBUG oslo_concurrency.lockutils [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.103559] env[69171]: DEBUG oslo_concurrency.lockutils [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.103647] env[69171]: DEBUG nova.compute.manager [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] No waiting events found dispatching network-vif-plugged-536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 {{(pid=69171) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 684.103757] env[69171]: WARNING nova.compute.manager [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Received unexpected event network-vif-plugged-536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 for instance with vm_state building and task_state spawning. [ 684.103930] env[69171]: DEBUG nova.compute.manager [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Received event network-changed-536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 684.107210] env[69171]: DEBUG nova.compute.manager [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Refreshing instance network info cache due to event network-changed-536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0. {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11137}} [ 684.107581] env[69171]: DEBUG oslo_concurrency.lockutils [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] Acquiring lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.107644] env[69171]: DEBUG oslo_concurrency.lockutils [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] Acquired lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.107795] env[69171]: DEBUG nova.network.neutron [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Refreshing network info cache for port 536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2007}} [ 684.111881] env[69171]: INFO nova.compute.manager [None req-fe3a6945-9e2d-4a94-bc68-64bf651ec741 tempest-ServerDiagnosticsV248Test-883600288 tempest-ServerDiagnosticsV248Test-883600288-project-admin] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Retrieving diagnostics [ 684.112682] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ff17a3-8f81-4668-91ca-d11a9897b828 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.225980] env[69171]: DEBUG nova.compute.manager [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 684.251385] env[69171]: DEBUG oslo_vmware.api [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235906, 'name': PowerOnVM_Task, 'duration_secs': 0.657232} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.251605] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Powered on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 684.251950] env[69171]: INFO nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Took 10.39 seconds to spawn the instance on the hypervisor. [ 684.252033] env[69171]: DEBUG nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 684.252953] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7f7b677-87dc-4c7d-bf97-7fdb6904f9ce {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.275203] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.290678] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235907, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.744386} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.290939] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85/8b6a1136-9dd6-4a9f-836d-e6e97a1c2b85.vmdk to [datastore1] fa5997fa-817c-42b6-a7fa-b3b85a109b6d/fa5997fa-817c-42b6-a7fa-b3b85a109b6d.vmdk {{(pid=69171) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1424}} [ 684.291165] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Extending root virtual disk to 1048576 {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 684.291412] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4d546e6a-af13-4fa9-ae55-d35afe751b5c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.299513] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 684.299513] env[69171]: value = "task-4235908" [ 684.299513] env[69171]: _type = "Task" [ 684.299513] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.308902] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235908, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 684.341666] env[69171]: DEBUG oslo_concurrency.lockutils [None req-48dcf33a-b365-4242-b81a-fa51334d4b49 tempest-ServerAddressesNegativeTestJSON-1406816668 tempest-ServerAddressesNegativeTestJSON-1406816668-project-member] Lock "836c5a40-0325-49d5-90d2-bd86cb3748ab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.353s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.497268] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance e00336c6-bcf5-439e-8838-f324423f800d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 684.760319] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.773153] env[69171]: INFO nova.compute.manager [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Took 16.84 seconds to build instance. [ 684.783752] env[69171]: INFO nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: d8590a5d-0b74-46da-93e4-01cb4d37c7d1] Took 1.05 seconds to deallocate network for instance. [ 684.816024] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235908, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075338} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 684.816024] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Extended root virtual disk {{(pid=69171) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 684.816024] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0a618ee-8c9b-4880-b2bd-ac84ce31221f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.841894] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Reconfiguring VM instance instance-00000006 to attach disk [datastore1] fa5997fa-817c-42b6-a7fa-b3b85a109b6d/fa5997fa-817c-42b6-a7fa-b3b85a109b6d.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 684.842811] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a4a99a69-4936-4134-9f89-18c1cae4d5e2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.867226] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 684.867226] env[69171]: value = "task-4235909" [ 684.867226] env[69171]: _type = "Task" [ 684.867226] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 684.879394] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235909, 'name': ReconfigVM_Task} progress is 6%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.000981] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 62413635-6f32-4539-b9ee-6ebe50733cf1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.274784] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3f919129-2498-4737-baf8-9ddca650c57d tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.349s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 685.381138] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235909, 'name': ReconfigVM_Task, 'duration_secs': 0.427752} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.381571] env[69171]: DEBUG nova.virt.vmwareapi.volumeops [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Reconfigured VM instance instance-00000006 to attach disk [datastore1] fa5997fa-817c-42b6-a7fa-b3b85a109b6d/fa5997fa-817c-42b6-a7fa-b3b85a109b6d.vmdk or device None with type sparse {{(pid=69171) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 685.382390] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-42a590b7-1da3-428d-9a39-598d1fadef6a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.389367] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 685.389367] env[69171]: value = "task-4235910" [ 685.389367] env[69171]: _type = "Task" [ 685.389367] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.396115] env[69171]: DEBUG nova.network.neutron [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Updated VIF entry in instance network info cache for port 536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0. {{(pid=69171) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3482}} [ 685.396473] env[69171]: DEBUG nova.network.neutron [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Updating instance_info_cache with network_info: [{"id": "536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0", "address": "fa:16:3e:10:14:88", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.82", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap536e9ab7-6a", "ovs_interfaceid": "536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.405415] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235910, 'name': Rename_Task} progress is 5%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 685.511359] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance eb6e8ba7-7ebd-48ec-b536-7e0468011057 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 685.833647] env[69171]: INFO nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Deleted allocations for instance d8590a5d-0b74-46da-93e4-01cb4d37c7d1 [ 685.900190] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235910, 'name': Rename_Task, 'duration_secs': 0.375464} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 685.900942] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Powering on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1442}} [ 685.901438] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0b831ec3-cbcd-477b-bb62-afa6a788353b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.904830] env[69171]: DEBUG oslo_concurrency.lockutils [req-915b13e3-d066-4d79-ba38-822d55e034bf req-d041b36d-ffe4-4d5d-93a4-79dde1b62b73 service nova] Releasing lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.912323] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for the task: (returnval){ [ 685.912323] env[69171]: value = "task-4235911" [ 685.912323] env[69171]: _type = "Task" [ 685.912323] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 685.924413] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235911, 'name': PowerOnVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.020267] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance dffe1923-6324-415c-9e38-680e2b213dca has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 686.020661] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 686.020733] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2048MB phys_disk=200GB used_disk=8GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 686.064145] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 686.086895] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 686.087181] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 686.106643] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 686.130914] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 686.321078] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49d26122-612f-4067-9b89-b8fd77a14b67 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.342900] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ff65bd-99c2-4a2f-b5ce-ceac2e7c664f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.346099] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "d8590a5d-0b74-46da-93e4-01cb4d37c7d1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.016s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.377388] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330bb9a3-1f07-41e4-bf8b-d144c1fa724b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.386397] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb52d29-2d5e-4504-aba4-828fcd17c0a6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.409543] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 686.422194] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235911, 'name': PowerOnVM_Task} progress is 1%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.547210] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.547596] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.548221] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.548221] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.548674] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.553048] env[69171]: INFO nova.compute.manager [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Terminating instance [ 686.557589] env[69171]: DEBUG nova.compute.manager [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Start destroying the instance on the hypervisor. {{(pid=69171) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3151}} [ 686.557589] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Destroying instance {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 686.558593] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f18b2a5-0b5e-44f5-91a1-54e4a902024e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.573583] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Powering off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 686.573866] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afe60363-3098-44c4-a6a8-0a7a118e7e7f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.586682] env[69171]: DEBUG oslo_vmware.api [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 686.586682] env[69171]: value = "task-4235912" [ 686.586682] env[69171]: _type = "Task" [ 686.586682] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 686.602231] env[69171]: DEBUG oslo_vmware.api [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235912, 'name': PowerOffVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.860294] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Acquiring lock "b41adb4d-5a5d-4e48-a4ed-6163e9c40e75" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.860551] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Lock "b41adb4d-5a5d-4e48-a4ed-6163e9c40e75" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.936671] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235911, 'name': PowerOnVM_Task} progress is 64%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 686.942723] env[69171]: ERROR nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [req-4465f752-fed4-46f6-9f0e-ef53c4659031] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-4465f752-fed4-46f6-9f0e-ef53c4659031"}]} [ 686.943263] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 4.502s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.944244] env[69171]: ERROR nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Error updating resources for node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28.: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 686.944244] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.944244] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 686.944244] env[69171]: ERROR nova.compute.manager yield [ 686.944244] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 686.944244] env[69171]: ERROR nova.compute.manager self.set_inventory_for_provider( [ 686.944244] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 686.944244] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 686.944244] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-4465f752-fed4-46f6-9f0e-ef53c4659031"}]} [ 686.944244] env[69171]: ERROR nova.compute.manager [ 686.944244] env[69171]: ERROR nova.compute.manager During handling of the above exception, another exception occurred: [ 686.944244] env[69171]: ERROR nova.compute.manager [ 686.946188] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 10584, in _update_available_resource_for_node [ 686.946188] env[69171]: ERROR nova.compute.manager self.rt.update_available_resource(context, nodename, [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 935, in update_available_resource [ 686.946188] env[69171]: ERROR nova.compute.manager self._update_available_resource(context, resources, startup=startup) [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 686.946188] env[69171]: ERROR nova.compute.manager return f(*args, **kwargs) [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1066, in _update_available_resource [ 686.946188] env[69171]: ERROR nova.compute.manager self._update(context, cn, startup=startup) [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 686.946188] env[69171]: ERROR nova.compute.manager self._update_to_placement(context, compute_node, startup) [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 686.946188] env[69171]: ERROR nova.compute.manager return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 686.946188] env[69171]: ERROR nova.compute.manager return attempt.get(self._wrap_exception) [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 686.946188] env[69171]: ERROR nova.compute.manager six.reraise(self.value[0], self.value[1], self.value[2]) [ 686.946188] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 686.947081] env[69171]: ERROR nova.compute.manager raise value [ 686.947081] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 686.947081] env[69171]: ERROR nova.compute.manager attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 686.947081] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 686.947081] env[69171]: ERROR nova.compute.manager self.reportclient.update_from_provider_tree( [ 686.947081] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 686.947081] env[69171]: ERROR nova.compute.manager with catch_all(pd.uuid): [ 686.947081] env[69171]: ERROR nova.compute.manager File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 686.947081] env[69171]: ERROR nova.compute.manager self.gen.throw(typ, value, traceback) [ 686.947081] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 686.947081] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderSyncFailed() [ 686.947081] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 686.947081] env[69171]: ERROR nova.compute.manager [ 686.947081] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.752s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.947625] env[69171]: INFO nova.compute.claims [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 687.101866] env[69171]: DEBUG oslo_vmware.api [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235912, 'name': PowerOffVM_Task, 'duration_secs': 0.336703} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.102104] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Powered off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 687.102325] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Unregistering the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 687.102729] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-07940739-5728-4f8d-9f8b-5931b53a455a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.192138] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Unregistered the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 687.192480] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Deleting contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 687.192742] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Deleting the datastore file [datastore1] f5a71cca-1e32-4c3c-9bb8-83e634a864f7 {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 687.194336] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-241a568e-6ebd-489d-a52c-30ff39f5d5a1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.204979] env[69171]: DEBUG oslo_vmware.api [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for the task: (returnval){ [ 687.204979] env[69171]: value = "task-4235914" [ 687.204979] env[69171]: _type = "Task" [ 687.204979] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 687.216414] env[69171]: DEBUG oslo_vmware.api [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.366077] env[69171]: DEBUG nova.compute.manager [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 687.434419] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235911, 'name': PowerOnVM_Task} progress is 91%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.720965] env[69171]: DEBUG oslo_vmware.api [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235914, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 687.898567] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.930953] env[69171]: DEBUG oslo_vmware.api [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Task: {'id': task-4235911, 'name': PowerOnVM_Task, 'duration_secs': 1.541215} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 687.931090] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Powered on the VM {{(pid=69171) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1448}} [ 687.931407] env[69171]: INFO nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Took 11.75 seconds to spawn the instance on the hypervisor. [ 687.931637] env[69171]: DEBUG nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 687.932538] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e351753-480f-4e50-add8-989aafde6528 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.986126] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 688.007425] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 688.007425] env[69171]: DEBUG nova.compute.provider_tree [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 688.024493] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 688.053048] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 688.067332] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "f4387f56-2438-4373-bcf5-7aee060b0d8c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.067332] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "f4387f56-2438-4373-bcf5-7aee060b0d8c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.160438] env[69171]: DEBUG nova.compute.manager [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Stashing vm_state: active {{(pid=69171) _prep_resize /opt/stack/nova/nova/compute/manager.py:5625}} [ 688.223028] env[69171]: DEBUG oslo_vmware.api [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Task: {'id': task-4235914, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.523555} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 688.223028] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Deleted the datastore file {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 688.223028] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Deleted contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 688.223028] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Instance destroyed {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 688.223028] env[69171]: INFO nova.compute.manager [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Took 1.67 seconds to destroy the instance on the hypervisor. [ 688.223311] env[69171]: DEBUG oslo.service.loopingcall [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.227216] env[69171]: DEBUG nova.compute.manager [-] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 688.227216] env[69171]: DEBUG nova.network.neutron [-] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 688.457598] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-164223d1-7712-46fe-bbd9-c306380eff92 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.466732] env[69171]: INFO nova.compute.manager [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Took 19.45 seconds to build instance. [ 688.471060] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78b50c7-097c-4d83-87b9-b1a728786bf1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.514441] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d21e9ac-b701-427e-8c2d-7dce0ca56158 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.527633] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3e1a1f8-773d-4e60-99d6-d90ecfbe7818 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.544521] env[69171]: DEBUG nova.compute.provider_tree [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 688.570976] env[69171]: DEBUG nova.compute.manager [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 688.691467] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 688.968172] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dbf881ec-34db-4e1b-9e32-071a48eaf498 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.960s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.075804] env[69171]: ERROR nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [req-545c3f0e-f213-41cc-91cb-e024b51be972] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-545c3f0e-f213-41cc-91cb-e024b51be972"}]} [ 689.076582] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.132s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.077406] env[69171]: ERROR nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 689.077406] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Traceback (most recent call last): [ 689.077406] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 689.077406] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] yield [ 689.077406] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 689.077406] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] self.set_inventory_for_provider( [ 689.077406] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 689.077406] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-545c3f0e-f213-41cc-91cb-e024b51be972"}]} [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] During handling of the above exception, another exception occurred: [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Traceback (most recent call last): [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] with self.rt.instance_claim(context, instance, node, allocs, [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 689.078444] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] return f(*args, **kwargs) [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] self._update(elevated, cn) [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] self._update_to_placement(context, compute_node, startup) [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] return attempt.get(self._wrap_exception) [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] six.reraise(self.value[0], self.value[1], self.value[2]) [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] raise value [ 689.079749] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] self.reportclient.update_from_provider_tree( [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] with catch_all(pd.uuid): [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] self.gen.throw(typ, value, traceback) [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] raise exception.ResourceProviderSyncFailed() [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 689.080150] env[69171]: ERROR nova.compute.manager [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] [ 689.080430] env[69171]: DEBUG nova.compute.utils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 689.080430] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.868s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.081771] env[69171]: INFO nova.compute.claims [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.087195] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Build of instance 5d998742-aa79-4b5b-8448-16ae18fa7b35 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 689.087689] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 689.087966] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "refresh_cache-5d998742-aa79-4b5b-8448-16ae18fa7b35" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.088176] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquired lock "refresh_cache-5d998742-aa79-4b5b-8448-16ae18fa7b35" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.088395] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 689.108650] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.644198] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 689.710225] env[69171]: DEBUG nova.compute.manager [None req-0b43e954-d773-4e79-81fa-07581178dc97 tempest-ServerDiagnosticsTest-968276502 tempest-ServerDiagnosticsTest-968276502-project-admin] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 689.713448] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a41081f9-398c-4815-adab-53f98f2b5829 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.722499] env[69171]: INFO nova.compute.manager [None req-0b43e954-d773-4e79-81fa-07581178dc97 tempest-ServerDiagnosticsTest-968276502 tempest-ServerDiagnosticsTest-968276502-project-admin] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Retrieving diagnostics [ 689.723915] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6305f8b6-e2b7-4879-ba5b-720c7ac684cd {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.729291] env[69171]: DEBUG nova.network.neutron [-] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.882580] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.128455] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 690.166328] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 690.166635] env[69171]: DEBUG nova.compute.provider_tree [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 690.197677] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 690.235315] env[69171]: INFO nova.compute.manager [-] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Took 2.01 seconds to deallocate network for instance. [ 690.254119] env[69171]: DEBUG nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 690.290770] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.290770] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.389524] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Releasing lock "refresh_cache-5d998742-aa79-4b5b-8448-16ae18fa7b35" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.392767] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 690.392767] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 690.392767] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 690.489831] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 690.637847] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17422d9b-e2e3-44ef-88ff-604076d32c62 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.650434] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4ec851-a4e6-47ad-b3a9-4c7c0cad9476 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.657110] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Acquiring lock "dfb3061a-b4e7-4d8a-bd71-876c88ff32d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.657373] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Lock "dfb3061a-b4e7-4d8a-bd71-876c88ff32d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.699083] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447aea51-63e8-4e67-a585-34cebc259b08 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.708353] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49720723-6c31-4e5b-b80c-038b1403bd1b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.724596] env[69171]: DEBUG nova.compute.provider_tree [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 690.747438] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.795423] env[69171]: DEBUG nova.compute.manager [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 690.993151] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.198539] env[69171]: DEBUG nova.compute.manager [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 691.257047] env[69171]: ERROR nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [req-1123f4d9-9c34-48fe-bc34-9a37c3d67137] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1123f4d9-9c34-48fe-bc34-9a37c3d67137"}]} [ 691.258108] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.258236] env[69171]: ERROR nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 691.258236] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] Traceback (most recent call last): [ 691.258236] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 691.258236] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] yield [ 691.258236] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 691.258236] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] self.set_inventory_for_provider( [ 691.258236] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 691.258236] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1123f4d9-9c34-48fe-bc34-9a37c3d67137"}]} [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] During handling of the above exception, another exception occurred: [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] Traceback (most recent call last): [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] with self.rt.instance_claim(context, instance, node, allocs, [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 691.261267] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] return f(*args, **kwargs) [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] self._update(elevated, cn) [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] self._update_to_placement(context, compute_node, startup) [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] return attempt.get(self._wrap_exception) [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] raise value [ 691.261892] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] self.reportclient.update_from_provider_tree( [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] with catch_all(pd.uuid): [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] self.gen.throw(typ, value, traceback) [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] raise exception.ResourceProviderSyncFailed() [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 691.262374] env[69171]: ERROR nova.compute.manager [instance: e00336c6-bcf5-439e-8838-f324423f800d] [ 691.262649] env[69171]: DEBUG nova.compute.utils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 691.267514] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.911s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.269847] env[69171]: INFO nova.compute.claims [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 691.285578] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Build of instance e00336c6-bcf5-439e-8838-f324423f800d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 691.286389] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 691.288891] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquiring lock "refresh_cache-e00336c6-bcf5-439e-8838-f324423f800d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.288891] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Acquired lock "refresh_cache-e00336c6-bcf5-439e-8838-f324423f800d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.288891] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 691.343863] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.375698] env[69171]: DEBUG nova.compute.manager [req-32dfeda2-9675-4e6b-b8ab-ef11e6fadd16 req-cfb5989e-f748-44ee-a172-8763ec4a7073 service nova] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Received event network-vif-deleted-aad071eb-62cd-4e3d-b8fc-62ea5e458141 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 691.500395] env[69171]: INFO nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: 5d998742-aa79-4b5b-8448-16ae18fa7b35] Took 1.11 seconds to deallocate network for instance. [ 691.729017] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.847153] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.309321] env[69171]: DEBUG nova.scheduler.client.report [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 692.331449] env[69171]: DEBUG nova.scheduler.client.report [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 692.331748] env[69171]: DEBUG nova.compute.provider_tree [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 692.354162] env[69171]: DEBUG nova.scheduler.client.report [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 692.359456] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.372624] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.372860] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.373059] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.373227] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.374195] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.380739] env[69171]: INFO nova.compute.manager [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Terminating instance [ 692.383117] env[69171]: DEBUG nova.scheduler.client.report [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 692.389457] env[69171]: DEBUG nova.compute.manager [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Start destroying the instance on the hypervisor. {{(pid=69171) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3151}} [ 692.390017] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Destroying instance {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 692.390807] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8221582-2beb-4165-8972-987f4a49f64f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.402045] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Powering off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 692.402697] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-35b67071-a60b-46fc-a164-bef6b116c673 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.413227] env[69171]: DEBUG oslo_vmware.api [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 692.413227] env[69171]: value = "task-4235915" [ 692.413227] env[69171]: _type = "Task" [ 692.413227] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 692.423786] env[69171]: DEBUG oslo_vmware.api [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235915, 'name': PowerOffVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 692.558094] env[69171]: INFO nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Deleted allocations for instance 5d998742-aa79-4b5b-8448-16ae18fa7b35 [ 692.745667] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0912048-530e-4e23-8645-dbbb282ebe6a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.758806] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68e5bf7c-0034-48bd-b889-a2058126691f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.793298] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d74672c0-878f-4aae-9490-5d0fd0095084 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.801251] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8dd56c8-ec43-4ac7-8f18-9f977cc889e5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.822430] env[69171]: DEBUG nova.compute.provider_tree [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 692.865900] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Releasing lock "refresh_cache-e00336c6-bcf5-439e-8838-f324423f800d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 692.865900] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 692.865900] env[69171]: DEBUG nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 692.865900] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 692.920228] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 692.927548] env[69171]: DEBUG oslo_vmware.api [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235915, 'name': PowerOffVM_Task, 'duration_secs': 0.216162} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 692.927824] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Powered off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 692.928706] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Unregistering the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 692.928706] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-71277be9-8e71-4959-a563-b214e0056976 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.001520] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Unregistered the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 693.001520] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Deleting contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 693.001520] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Deleting the datastore file [datastore1] 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03 {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 693.001758] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf063c7b-2d37-452c-b2aa-9ad4c54bc230 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.013961] env[69171]: DEBUG oslo_vmware.api [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for the task: (returnval){ [ 693.013961] env[69171]: value = "task-4235917" [ 693.013961] env[69171]: _type = "Task" [ 693.013961] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.029468] env[69171]: DEBUG oslo_vmware.api [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235917, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.071823] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "5d998742-aa79-4b5b-8448-16ae18fa7b35" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.683s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.371361] env[69171]: ERROR nova.scheduler.client.report [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [req-fe67457d-1b2f-4b66-8851-04560dda6764] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fe67457d-1b2f-4b66-8851-04560dda6764"}]} [ 693.372103] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.105s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.376022] env[69171]: ERROR nova.compute.manager [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 693.376022] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Traceback (most recent call last): [ 693.376022] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 693.376022] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] yield [ 693.376022] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 693.376022] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] self.set_inventory_for_provider( [ 693.376022] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 693.376022] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fe67457d-1b2f-4b66-8851-04560dda6764"}]} [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] During handling of the above exception, another exception occurred: [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Traceback (most recent call last): [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] with self.rt.instance_claim(context, instance, node, allocs, [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 693.376475] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] return f(*args, **kwargs) [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] self._update(elevated, cn) [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] self._update_to_placement(context, compute_node, startup) [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] return attempt.get(self._wrap_exception) [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] six.reraise(self.value[0], self.value[1], self.value[2]) [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] raise value [ 693.376747] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] self.reportclient.update_from_provider_tree( [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] with catch_all(pd.uuid): [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] self.gen.throw(typ, value, traceback) [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] raise exception.ResourceProviderSyncFailed() [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 693.377329] env[69171]: ERROR nova.compute.manager [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] [ 693.377631] env[69171]: DEBUG nova.compute.utils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 693.377631] env[69171]: DEBUG nova.compute.manager [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Build of instance 62413635-6f32-4539-b9ee-6ebe50733cf1 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 693.377631] env[69171]: DEBUG nova.compute.manager [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 693.377631] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Acquiring lock "refresh_cache-62413635-6f32-4539-b9ee-6ebe50733cf1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.377776] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Acquired lock "refresh_cache-62413635-6f32-4539-b9ee-6ebe50733cf1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.377776] env[69171]: DEBUG nova.network.neutron [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 693.381136] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.284s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.382668] env[69171]: INFO nova.compute.claims [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.426130] env[69171]: DEBUG nova.network.neutron [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.528806] env[69171]: DEBUG oslo_vmware.api [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Task: {'id': task-4235917, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.159093} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 693.529648] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Deleted the datastore file {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 693.529886] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Deleted contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 693.530689] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Instance destroyed {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 693.530689] env[69171]: INFO nova.compute.manager [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Took 1.14 seconds to destroy the instance on the hypervisor. [ 693.530793] env[69171]: DEBUG oslo.service.loopingcall [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.530977] env[69171]: DEBUG nova.compute.manager [-] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 693.531085] env[69171]: DEBUG nova.network.neutron [-] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 693.650490] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Acquiring lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.650490] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.650490] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Acquiring lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.650490] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.650903] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.655587] env[69171]: INFO nova.compute.manager [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Terminating instance [ 693.658308] env[69171]: DEBUG nova.compute.manager [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Start destroying the instance on the hypervisor. {{(pid=69171) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3151}} [ 693.658560] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Destroying instance {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 693.659426] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-319ccd20-336b-43f5-8c13-6ed1a80d507b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.668126] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Powering off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 693.668449] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fab68f2a-f2ae-4b08-825b-e9292df825f5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.677373] env[69171]: DEBUG oslo_vmware.api [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Waiting for the task: (returnval){ [ 693.677373] env[69171]: value = "task-4235918" [ 693.677373] env[69171]: _type = "Task" [ 693.677373] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 693.690520] env[69171]: DEBUG oslo_vmware.api [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Task: {'id': task-4235918, 'name': PowerOffVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 693.839897] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Acquiring lock "7a8fcc85-7f53-46fd-ad72-faea95991df1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 693.840660] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Lock "7a8fcc85-7f53-46fd-ad72-faea95991df1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.929798] env[69171]: DEBUG nova.network.neutron [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 693.932418] env[69171]: INFO nova.compute.manager [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] [instance: e00336c6-bcf5-439e-8838-f324423f800d] Took 1.07 seconds to deallocate network for instance. [ 694.189376] env[69171]: DEBUG oslo_vmware.api [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Task: {'id': task-4235918, 'name': PowerOffVM_Task, 'duration_secs': 0.284012} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.189646] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Powered off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 694.189808] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Unregistering the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 694.190114] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-24418885-9edd-4fdb-8b45-bae024cb4ecc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.256366] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Unregistered the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 694.256887] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Deleting contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 694.257346] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Deleting the datastore file [datastore1] fa5997fa-817c-42b6-a7fa-b3b85a109b6d {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 694.257920] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d04daf76-671f-42ba-aaf0-b97d0e8a59c1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.278177] env[69171]: DEBUG oslo_vmware.api [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Waiting for the task: (returnval){ [ 694.278177] env[69171]: value = "task-4235920" [ 694.278177] env[69171]: _type = "Task" [ 694.278177] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 694.290723] env[69171]: DEBUG oslo_vmware.api [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Task: {'id': task-4235920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 694.343468] env[69171]: DEBUG nova.compute.manager [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 694.402753] env[69171]: DEBUG nova.network.neutron [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.422526] env[69171]: DEBUG nova.scheduler.client.report [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 694.446023] env[69171]: DEBUG nova.scheduler.client.report [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 694.446253] env[69171]: DEBUG nova.compute.provider_tree [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.462647] env[69171]: DEBUG nova.scheduler.client.report [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 694.497688] env[69171]: DEBUG nova.scheduler.client.report [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 694.792413] env[69171]: DEBUG oslo_vmware.api [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Task: {'id': task-4235920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.34225} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 694.792860] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Deleted the datastore file {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 694.793196] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Deleted contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 694.793477] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance destroyed {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 694.793780] env[69171]: INFO nova.compute.manager [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Took 1.14 seconds to destroy the instance on the hypervisor. [ 694.794279] env[69171]: DEBUG oslo.service.loopingcall [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.794595] env[69171]: DEBUG nova.compute.manager [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 694.794786] env[69171]: DEBUG nova.network.neutron [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 694.824323] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab70e9ae-e8b6-468a-a278-955d68c39e6e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.834451] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fba086df-b929-43cb-9f50-80a50c500dd3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.874554] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f17e4a0-72e3-4b6c-869e-5f64bc6ac33f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.889657] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4cdb43-3bbf-4c26-9808-d27a404655a7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.907392] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Releasing lock "refresh_cache-62413635-6f32-4539-b9ee-6ebe50733cf1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.910608] env[69171]: DEBUG nova.compute.manager [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 694.910608] env[69171]: DEBUG nova.compute.manager [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 694.910801] env[69171]: DEBUG nova.network.neutron [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 694.912978] env[69171]: DEBUG nova.compute.provider_tree [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 694.918718] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.970582] env[69171]: INFO nova.scheduler.client.report [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Deleted allocations for instance e00336c6-bcf5-439e-8838-f324423f800d [ 695.017224] env[69171]: DEBUG nova.network.neutron [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 695.459587] env[69171]: ERROR nova.scheduler.client.report [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [req-374d73f8-e665-46bb-bf13-7ceb3217c456] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-374d73f8-e665-46bb-bf13-7ceb3217c456"}]} [ 695.460836] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.460836] env[69171]: ERROR nova.compute.manager [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 695.460836] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Traceback (most recent call last): [ 695.460836] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 695.460836] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] yield [ 695.460836] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 695.460836] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] self.set_inventory_for_provider( [ 695.460836] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 695.460836] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-374d73f8-e665-46bb-bf13-7ceb3217c456"}]} [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] During handling of the above exception, another exception occurred: [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Traceback (most recent call last): [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] with self.rt.instance_claim(context, instance, node, allocs, [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 695.461188] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] return f(*args, **kwargs) [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] self._update(elevated, cn) [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] self._update_to_placement(context, compute_node, startup) [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] return attempt.get(self._wrap_exception) [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] six.reraise(self.value[0], self.value[1], self.value[2]) [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] raise value [ 695.461493] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] self.reportclient.update_from_provider_tree( [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] with catch_all(pd.uuid): [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] self.gen.throw(typ, value, traceback) [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] raise exception.ResourceProviderSyncFailed() [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 695.461843] env[69171]: ERROR nova.compute.manager [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] [ 695.462216] env[69171]: DEBUG nova.compute.utils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 695.463432] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.703s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.464629] env[69171]: INFO nova.compute.claims [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 695.468412] env[69171]: DEBUG nova.compute.manager [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Build of instance eb6e8ba7-7ebd-48ec-b536-7e0468011057 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 695.472769] env[69171]: DEBUG nova.compute.manager [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 695.473198] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Acquiring lock "refresh_cache-eb6e8ba7-7ebd-48ec-b536-7e0468011057" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 695.473603] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Acquired lock "refresh_cache-eb6e8ba7-7ebd-48ec-b536-7e0468011057" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 695.473817] env[69171]: DEBUG nova.network.neutron [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 695.489358] env[69171]: DEBUG oslo_concurrency.lockutils [None req-08208dc4-4256-46cc-bff8-c8938a6eb99f tempest-ListServersNegativeTestJSON-103081067 tempest-ListServersNegativeTestJSON-103081067-project-member] Lock "e00336c6-bcf5-439e-8838-f324423f800d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.043s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.522578] env[69171]: DEBUG nova.network.neutron [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.685302] env[69171]: DEBUG nova.network.neutron [-] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.893513] env[69171]: DEBUG nova.compute.manager [None req-8aba5d5c-4a50-4803-8950-20fa391f818d tempest-ServerDiagnosticsV248Test-883600288 tempest-ServerDiagnosticsV248Test-883600288-project-admin] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Checking state {{(pid=69171) _get_power_state /opt/stack/nova/nova/compute/manager.py:1792}} [ 695.896875] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-478c845c-2a21-4ac5-989a-6af9c6e03036 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.905918] env[69171]: INFO nova.compute.manager [None req-8aba5d5c-4a50-4803-8950-20fa391f818d tempest-ServerDiagnosticsV248Test-883600288 tempest-ServerDiagnosticsV248Test-883600288-project-admin] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Retrieving diagnostics [ 695.907027] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72304401-75b7-4d44-ac38-62aafb542acd {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.021354] env[69171]: DEBUG nova.network.neutron [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.027765] env[69171]: INFO nova.compute.manager [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] [instance: 62413635-6f32-4539-b9ee-6ebe50733cf1] Took 1.12 seconds to deallocate network for instance. [ 696.192027] env[69171]: INFO nova.compute.manager [-] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Took 2.66 seconds to deallocate network for instance. [ 696.234786] env[69171]: DEBUG nova.network.neutron [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.303247] env[69171]: DEBUG nova.network.neutron [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 696.431292] env[69171]: DEBUG nova.compute.manager [req-e632bd2b-99bf-408f-856e-7215e83db0b2 req-f55183ce-01d5-4f84-988c-1863f4ebec74 service nova] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Received event network-vif-deleted-d8a791ab-d322-465f-80d1-73b383b7205b {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 696.505515] env[69171]: DEBUG nova.scheduler.client.report [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 696.522864] env[69171]: DEBUG nova.scheduler.client.report [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 696.523351] env[69171]: DEBUG nova.compute.provider_tree [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 696.545173] env[69171]: DEBUG nova.scheduler.client.report [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 696.573601] env[69171]: DEBUG nova.scheduler.client.report [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 696.701729] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.741243] env[69171]: INFO nova.compute.manager [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Took 1.94 seconds to deallocate network for instance. [ 696.808255] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Releasing lock "refresh_cache-eb6e8ba7-7ebd-48ec-b536-7e0468011057" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.808255] env[69171]: DEBUG nova.compute.manager [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 696.808255] env[69171]: DEBUG nova.compute.manager [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 696.808255] env[69171]: DEBUG nova.network.neutron [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 696.847184] env[69171]: DEBUG nova.network.neutron [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 696.862728] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02e1442d-24cd-4984-830e-cdbdecc1e13c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.877196] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9ac6ee9-3fc6-4d21-bb47-b99ea1b65e82 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.919869] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-058100a7-e756-4c3b-bb03-e748b59edff9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.930108] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9073be5-2c80-45ba-9281-c79c6a4071d9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 696.946329] env[69171]: DEBUG nova.compute.provider_tree [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 697.085457] env[69171]: INFO nova.scheduler.client.report [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Deleted allocations for instance 62413635-6f32-4539-b9ee-6ebe50733cf1 [ 697.248657] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 697.352599] env[69171]: DEBUG nova.network.neutron [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 697.437654] env[69171]: DEBUG nova.compute.manager [req-442a490e-d47a-44f8-8bd8-a9269efda5a4 req-eabdcd58-335a-4cb6-afd5-d9015824004b service nova] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Received event network-vif-deleted-536e9ab7-6a5a-496f-ba7b-fd4e9ee9a4a0 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 697.483514] env[69171]: ERROR nova.scheduler.client.report [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [req-49fe428b-185e-4ad8-9b03-a43176e6b115] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-49fe428b-185e-4ad8-9b03-a43176e6b115"}]} [ 697.483983] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.021s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.484467] env[69171]: ERROR nova.compute.manager [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 697.484467] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] Traceback (most recent call last): [ 697.484467] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 697.484467] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] yield [ 697.484467] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 697.484467] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] self.set_inventory_for_provider( [ 697.484467] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 697.484467] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-49fe428b-185e-4ad8-9b03-a43176e6b115"}]} [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] During handling of the above exception, another exception occurred: [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] Traceback (most recent call last): [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] with self.rt.instance_claim(context, instance, node, allocs, [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 697.484692] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] return f(*args, **kwargs) [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] self._update(elevated, cn) [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] self._update_to_placement(context, compute_node, startup) [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] return attempt.get(self._wrap_exception) [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] six.reraise(self.value[0], self.value[1], self.value[2]) [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] raise value [ 697.485100] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] self.reportclient.update_from_provider_tree( [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] with catch_all(pd.uuid): [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] self.gen.throw(typ, value, traceback) [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] raise exception.ResourceProviderSyncFailed() [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 697.485570] env[69171]: ERROR nova.compute.manager [instance: dffe1923-6324-415c-9e38-680e2b213dca] [ 697.485915] env[69171]: DEBUG nova.compute.utils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 697.488694] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.590s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 697.492453] env[69171]: INFO nova.compute.claims [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 697.499198] env[69171]: DEBUG nova.compute.manager [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Build of instance dffe1923-6324-415c-9e38-680e2b213dca was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 697.499198] env[69171]: DEBUG nova.compute.manager [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 697.499198] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Acquiring lock "refresh_cache-dffe1923-6324-415c-9e38-680e2b213dca" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.499198] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Acquired lock "refresh_cache-dffe1923-6324-415c-9e38-680e2b213dca" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.499504] env[69171]: DEBUG nova.network.neutron [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 697.599716] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7f2c2376-6596-4b2e-a036-551cc5f8915f tempest-ServersV294TestFqdnHostnames-666491170 tempest-ServersV294TestFqdnHostnames-666491170-project-member] Lock "62413635-6f32-4539-b9ee-6ebe50733cf1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.706s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.855909] env[69171]: INFO nova.compute.manager [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] [instance: eb6e8ba7-7ebd-48ec-b536-7e0468011057] Took 1.05 seconds to deallocate network for instance. [ 698.233430] env[69171]: DEBUG nova.network.neutron [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 698.469429] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "6a29d495-c938-41ad-97b6-1331b6eff589" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.469429] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "6a29d495-c938-41ad-97b6-1331b6eff589" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.469429] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "6a29d495-c938-41ad-97b6-1331b6eff589-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.469429] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "6a29d495-c938-41ad-97b6-1331b6eff589-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.469583] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "6a29d495-c938-41ad-97b6-1331b6eff589-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.470303] env[69171]: INFO nova.compute.manager [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Terminating instance [ 698.475140] env[69171]: DEBUG nova.network.neutron [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.476651] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "refresh_cache-6a29d495-c938-41ad-97b6-1331b6eff589" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 698.477040] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquired lock "refresh_cache-6a29d495-c938-41ad-97b6-1331b6eff589" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 698.477400] env[69171]: DEBUG nova.network.neutron [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 698.535472] env[69171]: DEBUG nova.scheduler.client.report [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 698.554350] env[69171]: DEBUG nova.scheduler.client.report [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 698.554350] env[69171]: DEBUG nova.compute.provider_tree [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 698.568328] env[69171]: DEBUG nova.scheduler.client.report [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 698.593550] env[69171]: DEBUG nova.scheduler.client.report [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 698.851562] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba1d4d01-069e-4941-8a39-d96463d9cee1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.862668] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65057ce0-92cc-40c3-b4db-03f121399ec1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.904723] env[69171]: INFO nova.scheduler.client.report [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Deleted allocations for instance eb6e8ba7-7ebd-48ec-b536-7e0468011057 [ 698.910620] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b01b803c-b3ae-4014-95aa-06a76842594b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.924408] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c2e3e35-6ca1-46f7-b732-78b709867125 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 698.939512] env[69171]: DEBUG nova.compute.provider_tree [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 698.979941] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Releasing lock "refresh_cache-dffe1923-6324-415c-9e38-680e2b213dca" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 698.980054] env[69171]: DEBUG nova.compute.manager [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 698.980302] env[69171]: DEBUG nova.compute.manager [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 698.980456] env[69171]: DEBUG nova.network.neutron [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 699.023878] env[69171]: DEBUG nova.network.neutron [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.055599] env[69171]: DEBUG nova.network.neutron [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 699.165188] env[69171]: DEBUG nova.network.neutron [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.416775] env[69171]: DEBUG oslo_concurrency.lockutils [None req-37614bec-fb93-4e10-9855-6e13ddb2db10 tempest-ImagesNegativeTestJSON-1685407278 tempest-ImagesNegativeTestJSON-1685407278-project-member] Lock "eb6e8ba7-7ebd-48ec-b536-7e0468011057" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.352s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.474875] env[69171]: ERROR nova.scheduler.client.report [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [req-29d93e65-cd1a-4d8b-9b4f-4392d3580cd0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-29d93e65-cd1a-4d8b-9b4f-4392d3580cd0"}]} [ 699.475036] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.987s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 699.475599] env[69171]: ERROR nova.compute.manager [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 699.475599] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Traceback (most recent call last): [ 699.475599] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 699.475599] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] yield [ 699.475599] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 699.475599] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] self.set_inventory_for_provider( [ 699.475599] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 699.475599] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-29d93e65-cd1a-4d8b-9b4f-4392d3580cd0"}]} [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] During handling of the above exception, another exception occurred: [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Traceback (most recent call last): [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] with self.rt.instance_claim(context, instance, node, allocs, [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 699.475868] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] return f(*args, **kwargs) [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] self._update(elevated, cn) [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] self._update_to_placement(context, compute_node, startup) [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] return attempt.get(self._wrap_exception) [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] six.reraise(self.value[0], self.value[1], self.value[2]) [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] raise value [ 699.476156] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] self.reportclient.update_from_provider_tree( [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] with catch_all(pd.uuid): [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] self.gen.throw(typ, value, traceback) [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] raise exception.ResourceProviderSyncFailed() [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 699.476502] env[69171]: ERROR nova.compute.manager [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] [ 699.476927] env[69171]: DEBUG nova.compute.utils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 699.477983] env[69171]: DEBUG nova.compute.manager [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Build of instance b41adb4d-5a5d-4e48-a4ed-6163e9c40e75 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 699.478395] env[69171]: DEBUG nova.compute.manager [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 699.478620] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Acquiring lock "refresh_cache-b41adb4d-5a5d-4e48-a4ed-6163e9c40e75" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.478796] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Acquired lock "refresh_cache-b41adb4d-5a5d-4e48-a4ed-6163e9c40e75" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.478965] env[69171]: DEBUG nova.network.neutron [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 699.480980] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 10.789s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.561602] env[69171]: DEBUG nova.network.neutron [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 699.668618] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Releasing lock "refresh_cache-6a29d495-c938-41ad-97b6-1331b6eff589" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.669074] env[69171]: DEBUG nova.compute.manager [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Start destroying the instance on the hypervisor. {{(pid=69171) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3151}} [ 699.669312] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Destroying instance {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 699.670907] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c4bac5-d55d-4245-88bd-8de4c0626a6f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.682385] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Powering off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 699.686343] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-203545be-229a-4892-8db8-85566e5cd73d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.694027] env[69171]: DEBUG oslo_vmware.api [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for the task: (returnval){ [ 699.694027] env[69171]: value = "task-4235921" [ 699.694027] env[69171]: _type = "Task" [ 699.694027] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 699.707771] env[69171]: DEBUG oslo_vmware.api [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Task: {'id': task-4235921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 699.730913] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquiring lock "8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 699.731221] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 699.986216] env[69171]: INFO nova.compute.claims [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 700.064216] env[69171]: INFO nova.compute.manager [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] [instance: dffe1923-6324-415c-9e38-680e2b213dca] Took 1.08 seconds to deallocate network for instance. [ 700.100264] env[69171]: DEBUG nova.network.neutron [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.215904] env[69171]: DEBUG oslo_vmware.exceptions [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Fault QuestionPending not matched. {{(pid=69171) get_fault_class /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/exceptions.py:290}} [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Destroy instance failed: oslo_vmware.exceptions.VimFaultException: The operation cannot be allowed at the current time because the virtual machine has a question pending: [ 700.217972] env[69171]: 'msg.hbacommon.outofspace:There is no more space for virtual disk '6a29d495-c938-41ad-97b6-1331b6eff589.vmdk'. You might be able to continue this session by freeing disk space on the relevant volume, and clicking Retry. Click Cancel to terminate this session. [ 700.217972] env[69171]: '. [ 700.217972] env[69171]: Faults: ['QuestionPending'] [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Traceback (most recent call last): [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1090, in _destroy_instance [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] vm_util.power_off_instance(self._session, instance, vm_ref) [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1506, in power_off_instance [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] session._wait_for_task(poweroff_task) [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] return self.wait_for_task(task_ref) [ 700.217972] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 398, in wait_for_task [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] return evt.wait() [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] result = hub.switch() [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] return self.greenlet.switch() [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self.f(*self.args, **self.kw) [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 448, in _poll_task [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] raise exceptions.translate_fault(task_info.error) [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] oslo_vmware.exceptions.VimFaultException: The operation cannot be allowed at the current time because the virtual machine has a question pending: [ 700.219599] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] 'msg.hbacommon.outofspace:There is no more space for virtual disk '6a29d495-c938-41ad-97b6-1331b6eff589.vmdk'. You might be able to continue this session by freeing disk space on the relevant volume, and clicking Retry. Click Cancel to terminate this session. [ 700.220277] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] '. [ 700.220277] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Faults: ['QuestionPending'] [ 700.220277] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] [ 700.220277] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Instance destroyed {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 700.222143] env[69171]: INFO nova.compute.manager [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Took 0.55 seconds to destroy the instance on the hypervisor. [ 700.222749] env[69171]: DEBUG oslo.service.loopingcall [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 700.222749] env[69171]: DEBUG nova.compute.manager [-] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 700.222892] env[69171]: DEBUG nova.network.neutron [-] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 700.236559] env[69171]: DEBUG nova.compute.manager [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 700.281796] env[69171]: DEBUG nova.network.neutron [-] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 700.493723] env[69171]: INFO nova.compute.resource_tracker [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updating resource usage from migration 7bd57b7a-be81-4c69-92f8-50364ba09d44 [ 700.516602] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 700.523069] env[69171]: DEBUG nova.network.neutron [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.531966] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 700.532454] env[69171]: DEBUG nova.compute.provider_tree [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 700.552130] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 700.592110] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 700.769934] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 700.787332] env[69171]: DEBUG nova.network.neutron [-] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.884380] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18ae3f2b-72b3-49fd-a6ea-cc13ac485893 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.891352] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91bd82a-be17-4c2e-971e-24fb1c2e4dc3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.928665] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24131595-c9cd-47c3-8b10-88051db1b09b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.937392] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afcba3fd-3947-4c44-b07a-007775e79549 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.952371] env[69171]: DEBUG nova.compute.provider_tree [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 701.026249] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Releasing lock "refresh_cache-b41adb4d-5a5d-4e48-a4ed-6163e9c40e75" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 701.026586] env[69171]: DEBUG nova.compute.manager [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 701.026637] env[69171]: DEBUG nova.compute.manager [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 701.026784] env[69171]: DEBUG nova.network.neutron [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 701.060430] env[69171]: DEBUG nova.network.neutron [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 701.119727] env[69171]: INFO nova.scheduler.client.report [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Deleted allocations for instance dffe1923-6324-415c-9e38-680e2b213dca [ 701.291459] env[69171]: INFO nova.compute.manager [-] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Took 1.07 seconds to deallocate network for instance. [ 701.486122] env[69171]: ERROR nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [req-198b453c-bcf0-4250-b85a-4091d3202998] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-198b453c-bcf0-4250-b85a-4091d3202998"}]} [ 701.489804] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.006s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.491955] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.381s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 701.495119] env[69171]: INFO nova.compute.claims [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 701.527044] env[69171]: INFO nova.compute.manager [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Swapping old allocation on dict_keys(['560320dd-0035-4785-a81a-5134ce1f7aaa']) held by migration 7bd57b7a-be81-4c69-92f8-50364ba09d44 for instance [ 701.563789] env[69171]: DEBUG nova.network.neutron [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.588606] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Overwriting current allocation {'allocations': {'560320dd-0035-4785-a81a-5134ce1f7aaa': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 28}}, 'project_id': '914e6c234dec4422a7448f447f80b588', 'user_id': '0b3d9fb8d4ba4df393db8e93ec169b86', 'consumer_generation': 1} on consumer f8b20421-15d3-46f3-a504-172d044ff4d3 {{(pid=69171) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 701.636636] env[69171]: DEBUG oslo_concurrency.lockutils [None req-eae2e7ba-a94e-46aa-a2f7-f83d41f6f87b tempest-InstanceActionsV221TestJSON-362765880 tempest-InstanceActionsV221TestJSON-362765880-project-member] Lock "dffe1923-6324-415c-9e38-680e2b213dca" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.916s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 701.697374] env[69171]: DEBUG nova.compute.manager [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Rescheduling, attempt 1 {{(pid=69171) _reschedule_resize_or_reraise /opt/stack/nova/nova/compute/manager.py:5789}} [ 701.802956] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.839948] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquiring lock "dfcf5a0c-1378-4dc6-8627-14a84919ce10" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 701.840194] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "dfcf5a0c-1378-4dc6-8627-14a84919ce10" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.068800] env[69171]: INFO nova.compute.manager [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] [instance: b41adb4d-5a5d-4e48-a4ed-6163e9c40e75] Took 1.04 seconds to deallocate network for instance. [ 702.208719] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.342924] env[69171]: DEBUG nova.compute.manager [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 702.541784] env[69171]: DEBUG nova.scheduler.client.report [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 702.566035] env[69171]: DEBUG nova.scheduler.client.report [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 702.566400] env[69171]: DEBUG nova.compute.provider_tree [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 702.587592] env[69171]: DEBUG nova.scheduler.client.report [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 702.630073] env[69171]: DEBUG nova.scheduler.client.report [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 702.878908] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.921128] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f213908-3c47-4a5e-bf16-ce40645cf1ed {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.931833] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57f322d5-b406-4ba9-b903-e7142f4986d0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.965363] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb173c3f-778d-4180-a8fa-74e8ad6d848f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.976524] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ba2268-59fc-4bd3-92fd-c3923961afad {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.996551] env[69171]: DEBUG nova.compute.provider_tree [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 703.135779] env[69171]: INFO nova.scheduler.client.report [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Deleted allocations for instance b41adb4d-5a5d-4e48-a4ed-6163e9c40e75 [ 703.368085] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquiring lock "56faf71b-84c3-44a0-b645-49a0849e2e02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 703.368351] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "56faf71b-84c3-44a0-b645-49a0849e2e02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.528995] env[69171]: ERROR nova.scheduler.client.report [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [req-fa7c279b-d8c9-45a1-85a1-461c229a06ef] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fa7c279b-d8c9-45a1-85a1-461c229a06ef"}]} [ 703.529322] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.040s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.530164] env[69171]: ERROR nova.compute.manager [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 703.530164] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Traceback (most recent call last): [ 703.530164] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 703.530164] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] yield [ 703.530164] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 703.530164] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] self.set_inventory_for_provider( [ 703.530164] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 703.530164] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fa7c279b-d8c9-45a1-85a1-461c229a06ef"}]} [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] During handling of the above exception, another exception occurred: [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Traceback (most recent call last): [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] with self.rt.instance_claim(context, instance, node, allocs, [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 703.530476] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] return f(*args, **kwargs) [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] self._update(elevated, cn) [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] self._update_to_placement(context, compute_node, startup) [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] return attempt.get(self._wrap_exception) [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] six.reraise(self.value[0], self.value[1], self.value[2]) [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] raise value [ 703.530758] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] self.reportclient.update_from_provider_tree( [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] with catch_all(pd.uuid): [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] self.gen.throw(typ, value, traceback) [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] raise exception.ResourceProviderSyncFailed() [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 703.531226] env[69171]: ERROR nova.compute.manager [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] [ 703.533608] env[69171]: DEBUG nova.compute.utils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 703.536156] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.793s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 703.536156] env[69171]: DEBUG nova.objects.instance [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lazy-loading 'resources' on Instance uuid f5a71cca-1e32-4c3c-9bb8-83e634a864f7 {{(pid=69171) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 703.536333] env[69171]: DEBUG nova.compute.manager [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Build of instance f4387f56-2438-4373-bcf5-7aee060b0d8c was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 703.540020] env[69171]: DEBUG nova.compute.manager [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 703.540020] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "refresh_cache-f4387f56-2438-4373-bcf5-7aee060b0d8c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.540020] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquired lock "refresh_cache-f4387f56-2438-4373-bcf5-7aee060b0d8c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.540020] env[69171]: DEBUG nova.network.neutron [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 703.653387] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b3a206e7-9461-4e0d-ad9c-46b0d96dd794 tempest-VolumesAssistedSnapshotsTest-1855500573 tempest-VolumesAssistedSnapshotsTest-1855500573-project-member] Lock "b41adb4d-5a5d-4e48-a4ed-6163e9c40e75" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.791s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 703.875540] env[69171]: DEBUG nova.compute.manager [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 704.077404] env[69171]: DEBUG nova.scheduler.client.report [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 704.107093] env[69171]: DEBUG nova.scheduler.client.report [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 704.109620] env[69171]: DEBUG nova.compute.provider_tree [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 704.141568] env[69171]: DEBUG nova.scheduler.client.report [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 704.188259] env[69171]: DEBUG nova.network.neutron [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 704.195226] env[69171]: DEBUG nova.scheduler.client.report [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 704.380055] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquiring lock "4b5077a6-2877-4e72-a85f-99a76d1a159c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.380230] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "4b5077a6-2877-4e72-a85f-99a76d1a159c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 704.404229] env[69171]: DEBUG nova.network.neutron [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.407218] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 704.490851] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-336d49b1-834f-4fd9-9bd7-ec08adf39eb9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.501309] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72819b98-9a5b-482c-b1cd-3b0b36c70d5a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.542641] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7e716d-6ede-43e4-9a6f-7d7668839444 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.553520] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795720f9-ca9b-49d8-9bd7-859066b0760c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.569572] env[69171]: DEBUG nova.compute.provider_tree [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 704.884598] env[69171]: DEBUG nova.compute.manager [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 704.907082] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Releasing lock "refresh_cache-f4387f56-2438-4373-bcf5-7aee060b0d8c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.907520] env[69171]: DEBUG nova.compute.manager [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 704.907605] env[69171]: DEBUG nova.compute.manager [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 704.907752] env[69171]: DEBUG nova.network.neutron [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 704.946130] env[69171]: DEBUG nova.network.neutron [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 705.099873] env[69171]: ERROR nova.scheduler.client.report [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [req-03c0e13b-9d0f-4c62-a115-c8b30e1ce1a3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-03c0e13b-9d0f-4c62-a115-c8b30e1ce1a3"}]} [ 705.100058] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.565s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.100888] env[69171]: ERROR nova.compute.manager [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 705.100888] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Traceback (most recent call last): [ 705.100888] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 705.100888] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] yield [ 705.100888] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 705.100888] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self.set_inventory_for_provider( [ 705.100888] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 705.100888] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-03c0e13b-9d0f-4c62-a115-c8b30e1ce1a3"}]} [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] During handling of the above exception, another exception occurred: [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Traceback (most recent call last): [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self._delete_instance(context, instance, bdms) [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 705.101206] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self._complete_deletion(context, instance) [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self._update_resource_tracker(context, instance) [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self.rt.update_usage(context, instance, instance.node) [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] return f(*args, **kwargs) [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self._update(context.elevated(), self.compute_nodes[nodename]) [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self._update_to_placement(context, compute_node, startup) [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 705.101605] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] return attempt.get(self._wrap_exception) [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] six.reraise(self.value[0], self.value[1], self.value[2]) [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] raise value [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self.reportclient.update_from_provider_tree( [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] with catch_all(pd.uuid): [ 705.102309] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 705.102823] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] self.gen.throw(typ, value, traceback) [ 705.102823] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 705.102823] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] raise exception.ResourceProviderSyncFailed() [ 705.102823] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 705.102823] env[69171]: ERROR nova.compute.manager [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] [ 705.104484] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.760s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.109716] env[69171]: INFO nova.compute.claims [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 705.411423] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.452157] env[69171]: DEBUG nova.network.neutron [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 705.618082] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "f5a71cca-1e32-4c3c-9bb8-83e634a864f7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.069s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.819833] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Acquiring lock "9f311bf0-31a1-46a9-911d-4e393c2279be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.820081] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Lock "9f311bf0-31a1-46a9-911d-4e393c2279be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.955518] env[69171]: INFO nova.compute.manager [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: f4387f56-2438-4373-bcf5-7aee060b0d8c] Took 1.05 seconds to deallocate network for instance. [ 706.159989] env[69171]: DEBUG nova.scheduler.client.report [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 706.189884] env[69171]: DEBUG nova.scheduler.client.report [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 706.190030] env[69171]: DEBUG nova.compute.provider_tree [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 706.221165] env[69171]: DEBUG nova.scheduler.client.report [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 706.253624] env[69171]: DEBUG nova.scheduler.client.report [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 706.324235] env[69171]: DEBUG nova.compute.manager [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 706.554671] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-447c0fb4-78c8-4813-b296-1b67d50da550 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.564324] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-458c9982-f85f-49fa-97d1-b3deba29eb7a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.605488] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-602f7dca-1679-4abf-b901-fecd931381b3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.614946] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2252b215-759b-452f-b6b1-7ae9ef633f49 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.634621] env[69171]: DEBUG nova.compute.provider_tree [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 706.853348] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.011515] env[69171]: INFO nova.scheduler.client.report [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Deleted allocations for instance f4387f56-2438-4373-bcf5-7aee060b0d8c [ 707.141783] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.175410] env[69171]: ERROR nova.scheduler.client.report [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [req-6986e60a-01bd-46c1-ba75-1734436bcde3] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6986e60a-01bd-46c1-ba75-1734436bcde3"}]} [ 707.175803] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.073s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.179153] env[69171]: ERROR nova.compute.manager [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 707.179153] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Traceback (most recent call last): [ 707.179153] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 707.179153] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] yield [ 707.179153] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 707.179153] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] self.set_inventory_for_provider( [ 707.179153] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 707.179153] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6986e60a-01bd-46c1-ba75-1734436bcde3"}]} [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] During handling of the above exception, another exception occurred: [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Traceback (most recent call last): [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] with self.rt.instance_claim(context, instance, node, allocs, [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 707.179415] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] return f(*args, **kwargs) [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] self._update(elevated, cn) [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] self._update_to_placement(context, compute_node, startup) [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] return attempt.get(self._wrap_exception) [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] raise value [ 707.179736] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] self.reportclient.update_from_provider_tree( [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] with catch_all(pd.uuid): [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] self.gen.throw(typ, value, traceback) [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] raise exception.ResourceProviderSyncFailed() [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 707.180105] env[69171]: ERROR nova.compute.manager [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] [ 707.180494] env[69171]: DEBUG nova.compute.utils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.183459] env[69171]: DEBUG nova.compute.manager [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Build of instance 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 707.183944] env[69171]: DEBUG nova.compute.manager [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 707.184225] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "refresh_cache-5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.184363] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquired lock "refresh_cache-5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.186805] env[69171]: DEBUG nova.network.neutron [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 707.186805] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.458s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.189226] env[69171]: INFO nova.compute.claims [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 707.523363] env[69171]: DEBUG oslo_concurrency.lockutils [None req-cc0694f0-fe47-49cc-81e1-69ef6ba01e2e tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "f4387f56-2438-4373-bcf5-7aee060b0d8c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.456s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.641794] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquiring lock "e1498893-a512-4b47-b22a-66c9e78df3de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.645254] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "e1498893-a512-4b47-b22a-66c9e78df3de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.739021] env[69171]: DEBUG nova.network.neutron [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 707.895644] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "d615fad6-4666-4a8d-a580-fc501abe501a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.895644] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "d615fad6-4666-4a8d-a580-fc501abe501a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 708.147315] env[69171]: DEBUG nova.compute.manager [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 708.229242] env[69171]: DEBUG nova.scheduler.client.report [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 708.247955] env[69171]: DEBUG nova.scheduler.client.report [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 708.248327] env[69171]: DEBUG nova.compute.provider_tree [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.262577] env[69171]: DEBUG nova.scheduler.client.report [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 708.294182] env[69171]: DEBUG nova.scheduler.client.report [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 708.302042] env[69171]: DEBUG nova.network.neutron [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.399925] env[69171]: DEBUG nova.compute.manager [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 708.654778] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e08fbf6a-18f3-4614-b4ba-620a888a9e28 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.674629] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b20a7e0f-8121-451a-b7ff-50b346bf4e8f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.716035] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 708.716412] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df23d164-0fa3-441a-bf76-9a78386695d6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.728602] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baef242d-31a3-48c7-8344-21afe91054bb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.745056] env[69171]: DEBUG nova.compute.provider_tree [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 708.804400] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Releasing lock "refresh_cache-5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.804654] env[69171]: DEBUG nova.compute.manager [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 708.804835] env[69171]: DEBUG nova.compute.manager [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 708.805099] env[69171]: DEBUG nova.network.neutron [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 708.890049] env[69171]: DEBUG nova.network.neutron [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 708.943603] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 709.278643] env[69171]: ERROR nova.scheduler.client.report [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [req-abec3b7e-bcda-4ca3-8689-5b86ad1d3f97] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-abec3b7e-bcda-4ca3-8689-5b86ad1d3f97"}]} [ 709.279182] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.092s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.279832] env[69171]: ERROR nova.compute.manager [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 709.279832] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Traceback (most recent call last): [ 709.279832] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 709.279832] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] yield [ 709.279832] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 709.279832] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] self.set_inventory_for_provider( [ 709.279832] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 709.279832] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-abec3b7e-bcda-4ca3-8689-5b86ad1d3f97"}]} [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] During handling of the above exception, another exception occurred: [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Traceback (most recent call last): [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] with self.rt.instance_claim(context, instance, node, allocs, [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 709.280119] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] return f(*args, **kwargs) [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] self._update(elevated, cn) [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] self._update_to_placement(context, compute_node, startup) [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] return attempt.get(self._wrap_exception) [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] six.reraise(self.value[0], self.value[1], self.value[2]) [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] raise value [ 709.280429] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] self.reportclient.update_from_provider_tree( [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] with catch_all(pd.uuid): [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] self.gen.throw(typ, value, traceback) [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] raise exception.ResourceProviderSyncFailed() [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 709.280817] env[69171]: ERROR nova.compute.manager [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] [ 709.282454] env[69171]: DEBUG nova.compute.utils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.283738] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.365s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.285718] env[69171]: INFO nova.compute.claims [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 709.290550] env[69171]: DEBUG nova.compute.manager [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Build of instance dfb3061a-b4e7-4d8a-bd71-876c88ff32d0 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 709.290857] env[69171]: DEBUG nova.compute.manager [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 709.291779] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Acquiring lock "refresh_cache-dfb3061a-b4e7-4d8a-bd71-876c88ff32d0" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.291779] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Acquired lock "refresh_cache-dfb3061a-b4e7-4d8a-bd71-876c88ff32d0" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.291972] env[69171]: DEBUG nova.network.neutron [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 709.397730] env[69171]: DEBUG nova.network.neutron [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.900221] env[69171]: INFO nova.compute.manager [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d] Took 1.09 seconds to deallocate network for instance. [ 709.908766] env[69171]: DEBUG nova.network.neutron [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 710.323507] env[69171]: DEBUG nova.scheduler.client.report [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 710.338752] env[69171]: DEBUG nova.scheduler.client.report [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 710.338970] env[69171]: DEBUG nova.compute.provider_tree [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 710.357094] env[69171]: DEBUG nova.scheduler.client.report [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 710.378554] env[69171]: DEBUG nova.scheduler.client.report [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 710.535805] env[69171]: DEBUG nova.network.neutron [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.788274] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Acquiring lock "f8c018fc-4d2e-4d71-ba99-7ca512c122df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.788274] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Lock "f8c018fc-4d2e-4d71-ba99-7ca512c122df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.789362] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c737e608-48c4-42dc-99f9-47a3a53976e4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.800847] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8749154b-319e-4c70-8f59-f3b8b2272cb4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.839789] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef777978-a76e-475e-abef-f380db8454f0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.849516] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee51dd05-69c8-41b3-8297-b40daae0c4be {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.867157] env[69171]: DEBUG nova.compute.provider_tree [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 710.958025] env[69171]: INFO nova.scheduler.client.report [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Deleted allocations for instance 5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d [ 711.042571] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Releasing lock "refresh_cache-dfb3061a-b4e7-4d8a-bd71-876c88ff32d0" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.044802] env[69171]: DEBUG nova.compute.manager [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 711.044802] env[69171]: DEBUG nova.compute.manager [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 711.044802] env[69171]: DEBUG nova.network.neutron [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 711.367020] env[69171]: DEBUG nova.network.neutron [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.396573] env[69171]: ERROR nova.scheduler.client.report [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [req-28030489-e876-4bf2-bc12-5728ef20eedd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-28030489-e876-4bf2-bc12-5728ef20eedd"}]} [ 711.396844] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.113s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.397499] env[69171]: ERROR nova.compute.manager [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 711.397499] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Traceback (most recent call last): [ 711.397499] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 711.397499] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] yield [ 711.397499] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 711.397499] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] self.set_inventory_for_provider( [ 711.397499] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 711.397499] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-28030489-e876-4bf2-bc12-5728ef20eedd"}]} [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] During handling of the above exception, another exception occurred: [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Traceback (most recent call last): [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] with self.rt.instance_claim(context, instance, node, allocs, [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 711.397827] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] return f(*args, **kwargs) [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] self._update(elevated, cn) [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] self._update_to_placement(context, compute_node, startup) [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] return attempt.get(self._wrap_exception) [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] six.reraise(self.value[0], self.value[1], self.value[2]) [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] raise value [ 711.398173] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] self.reportclient.update_from_provider_tree( [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] with catch_all(pd.uuid): [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] self.gen.throw(typ, value, traceback) [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] raise exception.ResourceProviderSyncFailed() [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 711.398517] env[69171]: ERROR nova.compute.manager [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] [ 711.399095] env[69171]: DEBUG nova.compute.utils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.400527] env[69171]: DEBUG nova.compute.manager [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Build of instance 7a8fcc85-7f53-46fd-ad72-faea95991df1 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 711.401111] env[69171]: DEBUG nova.compute.manager [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 711.402077] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Acquiring lock "refresh_cache-7a8fcc85-7f53-46fd-ad72-faea95991df1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.402374] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Acquired lock "refresh_cache-7a8fcc85-7f53-46fd-ad72-faea95991df1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.405219] env[69171]: DEBUG nova.network.neutron [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 711.406405] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.705s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.406580] env[69171]: DEBUG nova.objects.instance [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lazy-loading 'resources' on Instance uuid 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03 {{(pid=69171) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 711.475365] env[69171]: DEBUG oslo_concurrency.lockutils [None req-896a2e38-fc99-4a88-9709-6d77eeb8dfef tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "5d5ed70b-6de9-4fd9-8f22-10dc9aa1c28d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.187s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.872467] env[69171]: DEBUG nova.network.neutron [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 711.939027] env[69171]: DEBUG nova.scheduler.client.report [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 711.975187] env[69171]: DEBUG nova.scheduler.client.report [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 711.975187] env[69171]: DEBUG nova.compute.provider_tree [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 711.982035] env[69171]: DEBUG nova.compute.manager [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 711.987252] env[69171]: DEBUG nova.network.neutron [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 711.994590] env[69171]: DEBUG nova.scheduler.client.report [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 712.023059] env[69171]: DEBUG nova.scheduler.client.report [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 712.375125] env[69171]: INFO nova.compute.manager [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] [instance: dfb3061a-b4e7-4d8a-bd71-876c88ff32d0] Took 1.33 seconds to deallocate network for instance. [ 712.402638] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b514bc9e-9986-4788-bf01-d8a36f6c1d8d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.411739] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a89896cc-398a-4a4c-881d-ce07960efefc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.448196] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc20177-3a81-4a89-bbc5-759bed997f9e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.458299] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1edd80d4-d2f5-48e6-9b47-040ea0616b15 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.477077] env[69171]: DEBUG nova.compute.provider_tree [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 712.509292] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.607427] env[69171]: DEBUG nova.network.neutron [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.008797] env[69171]: ERROR nova.scheduler.client.report [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [req-011aea72-c900-4f8f-96a3-5563c122a635] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-011aea72-c900-4f8f-96a3-5563c122a635"}]} [ 713.009409] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.603s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.009817] env[69171]: ERROR nova.compute.manager [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 713.009817] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Traceback (most recent call last): [ 713.009817] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 713.009817] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] yield [ 713.009817] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 713.009817] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self.set_inventory_for_provider( [ 713.009817] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 713.009817] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-011aea72-c900-4f8f-96a3-5563c122a635"}]} [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] During handling of the above exception, another exception occurred: [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Traceback (most recent call last): [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self._delete_instance(context, instance, bdms) [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 713.010766] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self._complete_deletion(context, instance) [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self._update_resource_tracker(context, instance) [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self.rt.update_usage(context, instance, instance.node) [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] return f(*args, **kwargs) [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self._update(context.elevated(), self.compute_nodes[nodename]) [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self._update_to_placement(context, compute_node, startup) [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 713.011103] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] return attempt.get(self._wrap_exception) [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] six.reraise(self.value[0], self.value[1], self.value[2]) [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] raise value [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self.reportclient.update_from_provider_tree( [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] with catch_all(pd.uuid): [ 713.011501] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 713.011914] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] self.gen.throw(typ, value, traceback) [ 713.011914] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 713.011914] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] raise exception.ResourceProviderSyncFailed() [ 713.011914] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 713.011914] env[69171]: ERROR nova.compute.manager [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] [ 713.013837] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.765s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.013837] env[69171]: DEBUG nova.objects.instance [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lazy-loading 'resources' on Instance uuid fa5997fa-817c-42b6-a7fa-b3b85a109b6d {{(pid=69171) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 713.110063] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Releasing lock "refresh_cache-7a8fcc85-7f53-46fd-ad72-faea95991df1" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.110357] env[69171]: DEBUG nova.compute.manager [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 713.110671] env[69171]: DEBUG nova.compute.manager [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 713.110935] env[69171]: DEBUG nova.network.neutron [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 713.356222] env[69171]: DEBUG nova.network.neutron [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 713.419438] env[69171]: INFO nova.scheduler.client.report [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Deleted allocations for instance dfb3061a-b4e7-4d8a-bd71-876c88ff32d0 [ 713.518570] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.145s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.541125] env[69171]: DEBUG nova.scheduler.client.report [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 713.558937] env[69171]: DEBUG nova.scheduler.client.report [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 713.559204] env[69171]: DEBUG nova.compute.provider_tree [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 713.574568] env[69171]: DEBUG nova.scheduler.client.report [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 713.594615] env[69171]: DEBUG nova.scheduler.client.report [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 713.850378] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7025c3ce-5b37-46be-a479-e96a8cd06397 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.861738] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa02f1c1-8648-43fb-96b7-5be72e7a6df6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.866746] env[69171]: DEBUG nova.network.neutron [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.901979] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb61b11f-cd70-4da3-8b39-7e83596d17b6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.911350] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f46221cf-0404-4cee-8f68-f40c9d2f3269 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.931059] env[69171]: DEBUG nova.compute.provider_tree [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 713.936524] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07dc3be8-489a-4cd3-ab72-92e077aa8fce tempest-ServersWithSpecificFlavorTestJSON-321479728 tempest-ServersWithSpecificFlavorTestJSON-321479728-project-member] Lock "dfb3061a-b4e7-4d8a-bd71-876c88ff32d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.279s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.370669] env[69171]: INFO nova.compute.manager [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] [instance: 7a8fcc85-7f53-46fd-ad72-faea95991df1] Took 1.26 seconds to deallocate network for instance. [ 714.459545] env[69171]: ERROR nova.scheduler.client.report [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [req-73bdc089-765c-463f-8276-fb8c03a04fba] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-73bdc089-765c-463f-8276-fb8c03a04fba"}]} [ 714.459934] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.448s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.460647] env[69171]: ERROR nova.compute.manager [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 714.460647] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Traceback (most recent call last): [ 714.460647] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 714.460647] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] yield [ 714.460647] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 714.460647] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self.set_inventory_for_provider( [ 714.460647] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 714.460647] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-73bdc089-765c-463f-8276-fb8c03a04fba"}]} [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] During handling of the above exception, another exception occurred: [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Traceback (most recent call last): [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self._delete_instance(context, instance, bdms) [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 714.461047] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self._complete_deletion(context, instance) [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self._update_resource_tracker(context, instance) [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self.rt.update_usage(context, instance, instance.node) [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] return f(*args, **kwargs) [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self._update(context.elevated(), self.compute_nodes[nodename]) [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self._update_to_placement(context, compute_node, startup) [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 714.461348] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] return attempt.get(self._wrap_exception) [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] raise value [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self.reportclient.update_from_provider_tree( [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] with catch_all(pd.uuid): [ 714.461701] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 714.462184] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] self.gen.throw(typ, value, traceback) [ 714.462184] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 714.462184] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] raise exception.ResourceProviderSyncFailed() [ 714.462184] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 714.462184] env[69171]: ERROR nova.compute.manager [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] [ 714.464071] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.694s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.465783] env[69171]: INFO nova.compute.claims [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.971155] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.321s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.028347] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 715.418096] env[69171]: INFO nova.scheduler.client.report [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Deleted allocations for instance 7a8fcc85-7f53-46fd-ad72-faea95991df1 [ 715.507623] env[69171]: DEBUG nova.scheduler.client.report [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 715.529512] env[69171]: DEBUG nova.scheduler.client.report [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 715.529512] env[69171]: DEBUG nova.compute.provider_tree [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 715.541418] env[69171]: DEBUG nova.scheduler.client.report [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 715.576986] env[69171]: DEBUG nova.scheduler.client.report [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 715.873790] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de920018-e63c-4b6a-b302-6c6ea1aee87f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.883793] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-068a19a7-eb80-44d0-9125-d69676616b6f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.928332] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb742df-5fc2-492a-9393-c0f531e806c6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.931702] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f595f057-cc71-4470-9fe8-87dcdc637f20 tempest-ImagesOneServerTestJSON-1285312682 tempest-ImagesOneServerTestJSON-1285312682-project-member] Lock "7a8fcc85-7f53-46fd-ad72-faea95991df1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.091s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.940466] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f30ba4-3006-43dc-86e5-40e2f61a8123 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.957030] env[69171]: DEBUG nova.compute.provider_tree [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 716.490072] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.496120] env[69171]: ERROR nova.scheduler.client.report [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [req-71296201-5415-44c7-8937-3d4c50f60704] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-71296201-5415-44c7-8937-3d4c50f60704"}]} [ 716.496120] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.030s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 716.496679] env[69171]: ERROR nova.compute.manager [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 716.496679] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Traceback (most recent call last): [ 716.496679] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 716.496679] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] yield [ 716.496679] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 716.496679] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] self.set_inventory_for_provider( [ 716.496679] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 716.496679] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-71296201-5415-44c7-8937-3d4c50f60704"}]} [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] During handling of the above exception, another exception occurred: [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Traceback (most recent call last): [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] with self.rt.instance_claim(context, instance, node, allocs, [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 716.496907] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] return f(*args, **kwargs) [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] self._update(elevated, cn) [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] self._update_to_placement(context, compute_node, startup) [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] return attempt.get(self._wrap_exception) [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] raise value [ 716.497209] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] self.reportclient.update_from_provider_tree( [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] with catch_all(pd.uuid): [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] self.gen.throw(typ, value, traceback) [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] raise exception.ResourceProviderSyncFailed() [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 716.497643] env[69171]: ERROR nova.compute.manager [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] [ 716.497973] env[69171]: DEBUG nova.compute.utils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 716.500414] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.697s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.500664] env[69171]: DEBUG nova.objects.instance [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lazy-loading 'resources' on Instance uuid 6a29d495-c938-41ad-97b6-1331b6eff589 {{(pid=69171) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 716.505524] env[69171]: DEBUG nova.compute.manager [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Build of instance 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 716.505524] env[69171]: DEBUG nova.compute.manager [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 716.505524] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquiring lock "refresh_cache-8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.506116] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquired lock "refresh_cache-8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.506116] env[69171]: DEBUG nova.network.neutron [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 717.034020] env[69171]: DEBUG nova.scheduler.client.report [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 717.059773] env[69171]: DEBUG nova.scheduler.client.report [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 717.059773] env[69171]: DEBUG nova.compute.provider_tree [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 717.080294] env[69171]: DEBUG nova.network.neutron [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 717.092186] env[69171]: DEBUG nova.scheduler.client.report [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 717.122428] env[69171]: DEBUG nova.scheduler.client.report [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 717.439650] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9044ca04-179c-401f-9aa8-ff8bdb1cd929 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.450466] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b9d440-f7e8-429b-8e0e-76652776cea7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.495236] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471e55fe-455d-49b3-913c-e6a30567cd62 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.503433] env[69171]: DEBUG nova.network.neutron [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.507443] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Acquiring lock "4c596d23-7da7-4367-a60c-fe3b789516d9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.507443] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Lock "4c596d23-7da7-4367-a60c-fe3b789516d9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.525477] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45ab516a-66a6-4571-84d1-2a8cec717714 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.549032] env[69171]: DEBUG nova.compute.provider_tree [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 718.006075] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Releasing lock "refresh_cache-8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.006419] env[69171]: DEBUG nova.compute.manager [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 718.006643] env[69171]: DEBUG nova.compute.manager [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 718.007330] env[69171]: DEBUG nova.network.neutron [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 718.009674] env[69171]: DEBUG nova.compute.manager [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 718.087292] env[69171]: ERROR nova.scheduler.client.report [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [req-86f54cf9-68a7-4b09-89e2-3019b6f1f20f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-86f54cf9-68a7-4b09-89e2-3019b6f1f20f"}]} [ 718.087292] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.585s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.089612] env[69171]: ERROR nova.compute.manager [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 718.089612] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Traceback (most recent call last): [ 718.089612] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 718.089612] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] yield [ 718.089612] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 718.089612] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self.set_inventory_for_provider( [ 718.089612] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 718.089612] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-86f54cf9-68a7-4b09-89e2-3019b6f1f20f"}]} [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] During handling of the above exception, another exception occurred: [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Traceback (most recent call last): [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self._delete_instance(context, instance, bdms) [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 718.089923] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self._complete_deletion(context, instance) [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self._update_resource_tracker(context, instance) [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self.rt.update_usage(context, instance, instance.node) [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] return f(*args, **kwargs) [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self._update(context.elevated(), self.compute_nodes[nodename]) [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self._update_to_placement(context, compute_node, startup) [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 718.090225] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] return attempt.get(self._wrap_exception) [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] six.reraise(self.value[0], self.value[1], self.value[2]) [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] raise value [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self.reportclient.update_from_provider_tree( [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] with catch_all(pd.uuid): [ 718.091416] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 718.091980] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] self.gen.throw(typ, value, traceback) [ 718.091980] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 718.091980] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] raise exception.ResourceProviderSyncFailed() [ 718.091980] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 718.091980] env[69171]: ERROR nova.compute.manager [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] [ 718.094700] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.884s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.113889] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 718.123092] env[69171]: DEBUG nova.network.neutron [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 718.137497] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 718.137700] env[69171]: DEBUG nova.compute.provider_tree [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 718.158255] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 718.190615] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 718.537217] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.561534] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b416b9-1e9f-4c77-8a90-93a4be74a12b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.571327] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-883e15d9-cb1f-4519-a827-1edd2fb8ea69 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.618676] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "6a29d495-c938-41ad-97b6-1331b6eff589" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.149s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 718.618676] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8be4880-9a63-4d26-a59c-1d991d1999b3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.628756] env[69171]: DEBUG nova.network.neutron [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.631106] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dee2e826-9bdc-4dd4-a8bf-f46ade94a066 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.654301] env[69171]: DEBUG nova.compute.provider_tree [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 719.138852] env[69171]: INFO nova.compute.manager [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d] Took 1.13 seconds to deallocate network for instance. [ 719.183110] env[69171]: ERROR nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [req-9131d510-8531-4d56-9670-e5daa42a588d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9131d510-8531-4d56-9670-e5daa42a588d"}]}: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 719.183110] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.090s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.184303] env[69171]: ERROR nova.compute.manager [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Error trying to reschedule: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 719.184303] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Traceback (most recent call last): [ 719.184303] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 719.184303] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] yield [ 719.184303] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 719.184303] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.set_inventory_for_provider( [ 719.184303] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 719.184303] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-198b453c-bcf0-4250-b85a-4091d3202998"}]} [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] During handling of the above exception, another exception occurred: [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Traceback (most recent call last): [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 5749, in prep_resize [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._prep_resize(context, image, instance, [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 5662, in _prep_resize [ 719.184552] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] with self.rt.resize_claim( [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return f(*args, **kwargs) [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 238, in resize_claim [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return self._move_claim( [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 378, in _move_claim [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._update(elevated, cn) [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._update_to_placement(context, compute_node, startup) [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return attempt.get(self._wrap_exception) [ 719.184853] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise value [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.reportclient.update_from_provider_tree( [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] with catch_all(pd.uuid): [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.gen.throw(typ, value, traceback) [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 719.185266] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise exception.ResourceProviderSyncFailed() [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] During handling of the above exception, another exception occurred: [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Traceback (most recent call last): [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 5794, in _reschedule_resize_or_reraise [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._instance_update(context, instance, task_state=task_state) [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 701, in _instance_update [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._update_resource_tracker(context, instance) [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.rt.update_usage(context, instance, instance.node) [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 719.185721] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return f(*args, **kwargs) [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._update(context.elevated(), self.compute_nodes[nodename]) [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._update_to_placement(context, compute_node, startup) [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return attempt.get(self._wrap_exception) [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 719.186141] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise value [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.reportclient.update_from_provider_tree( [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] with catch_all(pd.uuid): [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.gen.throw(typ, value, traceback) [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise exception.ResourceProviderSyncFailed() [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 719.186527] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 719.189142] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.307s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.192563] env[69171]: INFO nova.compute.claims [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.701515] env[69171]: DEBUG nova.compute.utils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 720.096073] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "62f1950f-631d-4f07-8526-2155d4e1bc7c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.096491] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "62f1950f-631d-4f07-8526-2155d4e1bc7c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 720.135802] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 720.194201] env[69171]: INFO nova.scheduler.client.report [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Deleted allocations for instance 8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d [ 720.216199] env[69171]: ERROR nova.compute.manager [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 720.216199] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Traceback (most recent call last): [ 720.216199] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 720.216199] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] yield [ 720.216199] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 720.216199] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.set_inventory_for_provider( [ 720.216199] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 720.216199] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-198b453c-bcf0-4250-b85a-4091d3202998"}]} [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] During handling of the above exception, another exception occurred: [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Traceback (most recent call last): [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 10866, in _error_out_instance_on_exception [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] yield [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 5767, in prep_resize [ 720.216682] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._reschedule_resize_or_reraise(context, instance, [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 5847, in _reschedule_resize_or_reraise [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise exc [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 5749, in prep_resize [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._prep_resize(context, image, instance, [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/manager.py", line 5662, in _prep_resize [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] with self.rt.resize_claim( [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return f(*args, **kwargs) [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 238, in resize_claim [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return self._move_claim( [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 378, in _move_claim [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._update(elevated, cn) [ 720.217279] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self._update_to_placement(context, compute_node, startup) [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return attempt.get(self._wrap_exception) [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise value [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 720.218071] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.reportclient.update_from_provider_tree( [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] with catch_all(pd.uuid): [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.gen.throw(typ, value, traceback) [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise exception.ResourceProviderSyncFailed() [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 720.218648] env[69171]: ERROR nova.compute.manager [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 720.246016] env[69171]: DEBUG nova.scheduler.client.report [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 720.262299] env[69171]: DEBUG nova.scheduler.client.report [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 720.262620] env[69171]: DEBUG nova.compute.provider_tree [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 720.280296] env[69171]: DEBUG nova.scheduler.client.report [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: 186d9885-03e4-4156-9bff-74a5aaedd637 {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 720.305663] env[69171]: DEBUG nova.scheduler.client.report [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 720.604764] env[69171]: DEBUG nova.compute.manager [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 720.694528] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea4a6ab7-be38-421e-822a-57498b185f68 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.704338] env[69171]: DEBUG oslo_concurrency.lockutils [None req-43ce6db0-e119-47b7-901b-c2d354bd8386 tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "8e1d7fef-6ff1-4866-bcd0-9f46abb62a7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.973s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 720.706087] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f69bdfe-b59a-465a-877a-02d6307cf1b8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.754241] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d09ea04d-7945-42ab-b961-25209c8d6adf {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.764693] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb74bf6-9eb1-445e-abab-23918e416d35 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.780463] env[69171]: DEBUG nova.compute.provider_tree [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 721.144919] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.258391] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "bd91753e-aced-4c58-a6f1-f1baabbdd8c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.258391] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "bd91753e-aced-4c58-a6f1-f1baabbdd8c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.313237] env[69171]: ERROR nova.scheduler.client.report [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [req-e52620a5-ae03-4879-920d-2f86357351a0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e52620a5-ae03-4879-920d-2f86357351a0"}]} [ 721.318908] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.129s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.318908] env[69171]: ERROR nova.compute.manager [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 721.318908] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Traceback (most recent call last): [ 721.318908] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 721.318908] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] yield [ 721.318908] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 721.318908] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] self.set_inventory_for_provider( [ 721.318908] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 721.318908] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e52620a5-ae03-4879-920d-2f86357351a0"}]} [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] During handling of the above exception, another exception occurred: [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Traceback (most recent call last): [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] with self.rt.instance_claim(context, instance, node, allocs, [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 721.319242] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] return f(*args, **kwargs) [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] self._update(elevated, cn) [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] self._update_to_placement(context, compute_node, startup) [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] return attempt.get(self._wrap_exception) [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] six.reraise(self.value[0], self.value[1], self.value[2]) [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] raise value [ 721.319553] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] self.reportclient.update_from_provider_tree( [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] with catch_all(pd.uuid): [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] self.gen.throw(typ, value, traceback) [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] raise exception.ResourceProviderSyncFailed() [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 721.319962] env[69171]: ERROR nova.compute.manager [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] [ 721.320315] env[69171]: DEBUG nova.compute.utils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.320315] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.910s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.320315] env[69171]: INFO nova.compute.claims [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.325109] env[69171]: DEBUG nova.compute.manager [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Build of instance dfcf5a0c-1378-4dc6-8627-14a84919ce10 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 721.325109] env[69171]: DEBUG nova.compute.manager [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 721.325109] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquiring lock "refresh_cache-dfcf5a0c-1378-4dc6-8627-14a84919ce10" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.325109] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Acquired lock "refresh_cache-dfcf5a0c-1378-4dc6-8627-14a84919ce10" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.326271] env[69171]: DEBUG nova.network.neutron [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 721.767321] env[69171]: DEBUG nova.compute.manager [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 721.972219] env[69171]: DEBUG nova.network.neutron [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 722.270198] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.297513] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.352253] env[69171]: DEBUG nova.scheduler.client.report [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 722.372901] env[69171]: DEBUG nova.scheduler.client.report [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 722.373152] env[69171]: DEBUG nova.compute.provider_tree [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 722.390713] env[69171]: DEBUG nova.scheduler.client.report [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: 186d9885-03e4-4156-9bff-74a5aaedd637 {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 722.416286] env[69171]: DEBUG nova.scheduler.client.report [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 722.660668] env[69171]: DEBUG nova.network.neutron [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.751344] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdbc74ea-6eff-4c1f-9902-5e34083b1ef9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.765021] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1891e37b-ceda-4d76-94ea-8978db2f2745 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.804760] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ffa51e-ca39-4d90-88dc-20bd42cf36ff {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.813384] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ab245a-4819-4534-ac2c-719b4bef9e4b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.830525] env[69171]: DEBUG nova.compute.provider_tree [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 723.162952] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Releasing lock "refresh_cache-dfcf5a0c-1378-4dc6-8627-14a84919ce10" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 723.167196] env[69171]: DEBUG nova.compute.manager [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 723.167196] env[69171]: DEBUG nova.compute.manager [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 723.167196] env[69171]: DEBUG nova.network.neutron [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 723.442431] env[69171]: DEBUG nova.network.neutron [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 723.442431] env[69171]: ERROR nova.scheduler.client.report [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [req-f560c75e-a430-40aa-a126-1dbf85feb495] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f560c75e-a430-40aa-a126-1dbf85feb495"}]} [ 723.444301] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.063s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.444301] env[69171]: ERROR nova.compute.manager [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 723.444301] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Traceback (most recent call last): [ 723.444301] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 723.444301] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] yield [ 723.444301] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 723.444301] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] self.set_inventory_for_provider( [ 723.444301] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 723.444301] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f560c75e-a430-40aa-a126-1dbf85feb495"}]} [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] During handling of the above exception, another exception occurred: [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Traceback (most recent call last): [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] with self.rt.instance_claim(context, instance, node, allocs, [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 723.445520] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] return f(*args, **kwargs) [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] self._update(elevated, cn) [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] self._update_to_placement(context, compute_node, startup) [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] return attempt.get(self._wrap_exception) [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] six.reraise(self.value[0], self.value[1], self.value[2]) [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] raise value [ 723.446567] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] self.reportclient.update_from_provider_tree( [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] with catch_all(pd.uuid): [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] self.gen.throw(typ, value, traceback) [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] raise exception.ResourceProviderSyncFailed() [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 723.448526] env[69171]: ERROR nova.compute.manager [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] [ 723.449408] env[69171]: DEBUG nova.compute.utils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 723.449408] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.973s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.449408] env[69171]: INFO nova.compute.claims [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 723.449408] env[69171]: DEBUG nova.compute.manager [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Build of instance 56faf71b-84c3-44a0-b645-49a0849e2e02 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 723.449859] env[69171]: DEBUG nova.compute.manager [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 723.449859] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquiring lock "refresh_cache-56faf71b-84c3-44a0-b645-49a0849e2e02" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.449859] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquired lock "refresh_cache-56faf71b-84c3-44a0-b645-49a0849e2e02" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.449859] env[69171]: DEBUG nova.network.neutron [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 723.753277] env[69171]: DEBUG nova.network.neutron [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.774927] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Acquiring lock "f1a31770-1165-4143-be50-0e6a6ce2f041" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.775349] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Lock "f1a31770-1165-4143-be50-0e6a6ce2f041" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.999353] env[69171]: DEBUG nova.network.neutron [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 724.261092] env[69171]: INFO nova.compute.manager [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] [instance: dfcf5a0c-1378-4dc6-8627-14a84919ce10] Took 1.09 seconds to deallocate network for instance. [ 724.412675] env[69171]: DEBUG nova.network.neutron [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.434454] env[69171]: DEBUG nova.scheduler.client.report [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 724.462854] env[69171]: DEBUG nova.scheduler.client.report [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 724.462854] env[69171]: DEBUG nova.compute.provider_tree [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.488751] env[69171]: DEBUG nova.scheduler.client.report [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: 186d9885-03e4-4156-9bff-74a5aaedd637 {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 724.520999] env[69171]: DEBUG nova.scheduler.client.report [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 724.578248] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquiring lock "3b8a2f7d-51a2-42f1-ad30-12e3a1083037" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.578248] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "3b8a2f7d-51a2-42f1-ad30-12e3a1083037" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.743502] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Acquiring lock "363084be-ae28-433b-b7e0-d0a4eccbff6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.743502] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Lock "363084be-ae28-433b-b7e0-d0a4eccbff6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.890883] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34a9c705-fbf3-4aff-8d38-b17c28d49a9d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.899190] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bb4be56-316d-48c8-a8ab-0392d2de6114 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.934912] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Releasing lock "refresh_cache-56faf71b-84c3-44a0-b645-49a0849e2e02" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.935222] env[69171]: DEBUG nova.compute.manager [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 724.935427] env[69171]: DEBUG nova.compute.manager [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 724.935524] env[69171]: DEBUG nova.network.neutron [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 724.937957] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1353b55c-8193-40b3-bec3-cd9137028a07 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.946572] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926916dc-90b7-4830-9c9e-b197449feb57 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.961156] env[69171]: DEBUG nova.compute.provider_tree [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 724.987934] env[69171]: DEBUG nova.network.neutron [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 725.317874] env[69171]: INFO nova.scheduler.client.report [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Deleted allocations for instance dfcf5a0c-1378-4dc6-8627-14a84919ce10 [ 725.491063] env[69171]: DEBUG nova.network.neutron [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.502369] env[69171]: ERROR nova.scheduler.client.report [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [req-e4eaaad8-ffdd-4b08-b100-81684bcba7de] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e4eaaad8-ffdd-4b08-b100-81684bcba7de"}]} [ 725.502369] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.118s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.502632] env[69171]: ERROR nova.compute.manager [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 725.502632] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Traceback (most recent call last): [ 725.502632] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 725.502632] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] yield [ 725.502632] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 725.502632] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] self.set_inventory_for_provider( [ 725.502632] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 725.502632] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e4eaaad8-ffdd-4b08-b100-81684bcba7de"}]} [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] During handling of the above exception, another exception occurred: [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Traceback (most recent call last): [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] with self.rt.instance_claim(context, instance, node, allocs, [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 725.503138] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] return f(*args, **kwargs) [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] self._update(elevated, cn) [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] self._update_to_placement(context, compute_node, startup) [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] return attempt.get(self._wrap_exception) [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] six.reraise(self.value[0], self.value[1], self.value[2]) [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] raise value [ 725.503437] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] self.reportclient.update_from_provider_tree( [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] with catch_all(pd.uuid): [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] self.gen.throw(typ, value, traceback) [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] raise exception.ResourceProviderSyncFailed() [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 725.503805] env[69171]: ERROR nova.compute.manager [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] [ 725.504171] env[69171]: DEBUG nova.compute.utils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.506715] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.654s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.508260] env[69171]: INFO nova.compute.claims [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.516036] env[69171]: DEBUG nova.compute.manager [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Build of instance 4b5077a6-2877-4e72-a85f-99a76d1a159c was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 725.516036] env[69171]: DEBUG nova.compute.manager [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 725.516036] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquiring lock "refresh_cache-4b5077a6-2877-4e72-a85f-99a76d1a159c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.516036] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquired lock "refresh_cache-4b5077a6-2877-4e72-a85f-99a76d1a159c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.516313] env[69171]: DEBUG nova.network.neutron [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 725.830034] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b39bd1d2-a82e-40fe-915a-d5f53fa9e4bb tempest-ServersAdminTestJSON-204161610 tempest-ServersAdminTestJSON-204161610-project-member] Lock "dfcf5a0c-1378-4dc6-8627-14a84919ce10" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.990s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.996759] env[69171]: INFO nova.compute.manager [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: 56faf71b-84c3-44a0-b645-49a0849e2e02] Took 1.06 seconds to deallocate network for instance. [ 726.078379] env[69171]: DEBUG nova.network.neutron [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.334158] env[69171]: DEBUG nova.compute.manager [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 726.339643] env[69171]: DEBUG nova.network.neutron [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.556779] env[69171]: DEBUG nova.scheduler.client.report [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 726.582990] env[69171]: DEBUG nova.scheduler.client.report [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 726.582990] env[69171]: DEBUG nova.compute.provider_tree [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 726.600807] env[69171]: DEBUG nova.scheduler.client.report [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 726.631622] env[69171]: DEBUG nova.scheduler.client.report [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 726.847900] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Releasing lock "refresh_cache-4b5077a6-2877-4e72-a85f-99a76d1a159c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 726.847900] env[69171]: DEBUG nova.compute.manager [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 726.847900] env[69171]: DEBUG nova.compute.manager [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 726.847900] env[69171]: DEBUG nova.network.neutron [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 726.859980] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.896935] env[69171]: DEBUG nova.network.neutron [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 726.966238] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aac91df4-2f77-419b-9842-96448e8e03ce {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.975714] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7514df1-5748-4348-9663-b21de4b8e693 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.018141] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ddbb75-59d8-4c24-93e8-5e7c2e1dae75 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.027554] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09b0695d-304c-4c9c-8edd-e443f831d197 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.046891] env[69171]: DEBUG nova.compute.provider_tree [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 727.048777] env[69171]: INFO nova.scheduler.client.report [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Deleted allocations for instance 56faf71b-84c3-44a0-b645-49a0849e2e02 [ 727.404337] env[69171]: DEBUG nova.network.neutron [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.557402] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c206e1f5-8b0b-47bc-ada7-3c5c1d78f086 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "56faf71b-84c3-44a0-b645-49a0849e2e02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.189s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.587020] env[69171]: ERROR nova.scheduler.client.report [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [req-b5e0ab83-7068-40ce-8cc4-2a42a6837930] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b5e0ab83-7068-40ce-8cc4-2a42a6837930"}]} [ 727.587020] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.587225] env[69171]: ERROR nova.compute.manager [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 727.587225] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Traceback (most recent call last): [ 727.587225] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 727.587225] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] yield [ 727.587225] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 727.587225] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] self.set_inventory_for_provider( [ 727.587225] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 727.587225] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b5e0ab83-7068-40ce-8cc4-2a42a6837930"}]} [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] During handling of the above exception, another exception occurred: [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Traceback (most recent call last): [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] with self.rt.instance_claim(context, instance, node, allocs, [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 727.587507] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] return f(*args, **kwargs) [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] self._update(elevated, cn) [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] self._update_to_placement(context, compute_node, startup) [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] return attempt.get(self._wrap_exception) [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] six.reraise(self.value[0], self.value[1], self.value[2]) [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] raise value [ 727.587825] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] self.reportclient.update_from_provider_tree( [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] with catch_all(pd.uuid): [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] self.gen.throw(typ, value, traceback) [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] raise exception.ResourceProviderSyncFailed() [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 727.588233] env[69171]: ERROR nova.compute.manager [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] [ 727.588586] env[69171]: DEBUG nova.compute.utils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 727.589598] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.448s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.589820] env[69171]: DEBUG oslo_concurrency.lockutils [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 727.590258] env[69171]: INFO nova.compute.manager [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Successfully reverted task state from None on failure for instance. [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server [None req-abdd54f6-05a2-4443-b4ab-e3c6a716c303 tempest-ServerDiagnosticsNegativeTest-39838829 tempest-ServerDiagnosticsNegativeTest-39838829-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server yield [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-03c0e13b-9d0f-4c62-a115-c8b30e1ce1a3"}]} [ 727.599369] env[69171]: ERROR oslo_messaging.rpc.server [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 727.599739] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 727.600250] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3354, in terminate_instance [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3349, in do_terminate_instance [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 727.601025] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server raise value [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 727.601868] env[69171]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 727.602948] env[69171]: ERROR oslo_messaging.rpc.server [ 727.602948] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.882s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.602948] env[69171]: INFO nova.compute.claims [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 727.603515] env[69171]: DEBUG nova.compute.manager [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Build of instance 9f311bf0-31a1-46a9-911d-4e393c2279be was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 727.604208] env[69171]: DEBUG nova.compute.manager [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 727.604294] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Acquiring lock "refresh_cache-9f311bf0-31a1-46a9-911d-4e393c2279be" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.604474] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Acquired lock "refresh_cache-9f311bf0-31a1-46a9-911d-4e393c2279be" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.604672] env[69171]: DEBUG nova.network.neutron [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 727.817104] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Acquiring lock "e0f59163-b386-45d3-87d7-1271deebe2f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.817480] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Lock "e0f59163-b386-45d3-87d7-1271deebe2f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.888426] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "7a5215be-525e-4400-8972-4d6a949f5c24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.888714] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "7a5215be-525e-4400-8972-4d6a949f5c24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 727.905036] env[69171]: INFO nova.compute.manager [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: 4b5077a6-2877-4e72-a85f-99a76d1a159c] Took 1.06 seconds to deallocate network for instance. [ 727.916582] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "598fe41b-debf-4fbd-ac21-5cda132d2fea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.917654] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "598fe41b-debf-4fbd-ac21-5cda132d2fea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.063304] env[69171]: DEBUG nova.compute.manager [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 728.167927] env[69171]: DEBUG nova.network.neutron [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 728.493676] env[69171]: DEBUG nova.network.neutron [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.514222] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Acquiring lock "377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.514514] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Lock "377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.587797] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.637491] env[69171]: DEBUG nova.scheduler.client.report [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 728.652653] env[69171]: DEBUG nova.scheduler.client.report [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 728.652880] env[69171]: DEBUG nova.compute.provider_tree [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 728.665926] env[69171]: DEBUG nova.scheduler.client.report [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 728.684383] env[69171]: DEBUG nova.scheduler.client.report [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 728.945786] env[69171]: INFO nova.scheduler.client.report [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Deleted allocations for instance 4b5077a6-2877-4e72-a85f-99a76d1a159c [ 729.001849] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Releasing lock "refresh_cache-9f311bf0-31a1-46a9-911d-4e393c2279be" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.002161] env[69171]: DEBUG nova.compute.manager [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 729.002489] env[69171]: DEBUG nova.compute.manager [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 729.002578] env[69171]: DEBUG nova.network.neutron [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 729.028083] env[69171]: DEBUG nova.network.neutron [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 729.041410] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-358a6765-f840-4952-8507-4e0ef0f6f20e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.055922] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc59f953-238b-42ba-a8aa-2093b79c4968 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.097064] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d2b34c-fa17-4e94-9abe-90647710d150 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.107852] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93041bc0-a4e0-44cc-b6e2-1d39f457de71 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.128374] env[69171]: DEBUG nova.compute.provider_tree [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 729.436994] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Acquiring lock "d4220806-373f-45cc-82d0-cc70ca2c26ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.436994] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Lock "d4220806-373f-45cc-82d0-cc70ca2c26ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.460340] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a563c8ee-3410-4794-a4c9-5c0ac4d3179f tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "4b5077a6-2877-4e72-a85f-99a76d1a159c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.080s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.536984] env[69171]: DEBUG nova.network.neutron [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.662205] env[69171]: ERROR nova.scheduler.client.report [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [req-0f4ffd1a-a63e-4ec5-8f66-a1516525b694] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0f4ffd1a-a63e-4ec5-8f66-a1516525b694"}]} [ 729.662205] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.062s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.662401] env[69171]: ERROR nova.compute.manager [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 729.662401] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Traceback (most recent call last): [ 729.662401] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 729.662401] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] yield [ 729.662401] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 729.662401] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] self.set_inventory_for_provider( [ 729.662401] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 729.662401] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0f4ffd1a-a63e-4ec5-8f66-a1516525b694"}]} [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] During handling of the above exception, another exception occurred: [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Traceback (most recent call last): [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] with self.rt.instance_claim(context, instance, node, allocs, [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 729.662668] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] return f(*args, **kwargs) [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] self._update(elevated, cn) [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] self._update_to_placement(context, compute_node, startup) [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] return attempt.get(self._wrap_exception) [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] six.reraise(self.value[0], self.value[1], self.value[2]) [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] raise value [ 729.662995] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] self.reportclient.update_from_provider_tree( [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] with catch_all(pd.uuid): [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] self.gen.throw(typ, value, traceback) [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] raise exception.ResourceProviderSyncFailed() [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 729.663417] env[69171]: ERROR nova.compute.manager [instance: e1498893-a512-4b47-b22a-66c9e78df3de] [ 729.663751] env[69171]: DEBUG nova.compute.utils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 729.667518] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.724s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 729.668147] env[69171]: INFO nova.compute.claims [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 729.671345] env[69171]: DEBUG nova.compute.manager [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Build of instance e1498893-a512-4b47-b22a-66c9e78df3de was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 729.671783] env[69171]: DEBUG nova.compute.manager [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 729.672013] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquiring lock "refresh_cache-e1498893-a512-4b47-b22a-66c9e78df3de" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.672160] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Acquired lock "refresh_cache-e1498893-a512-4b47-b22a-66c9e78df3de" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.672311] env[69171]: DEBUG nova.network.neutron [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 729.966026] env[69171]: DEBUG nova.compute.manager [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 730.039328] env[69171]: INFO nova.compute.manager [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] [instance: 9f311bf0-31a1-46a9-911d-4e393c2279be] Took 1.04 seconds to deallocate network for instance. [ 730.227020] env[69171]: DEBUG nova.network.neutron [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 730.497378] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.531054] env[69171]: DEBUG nova.network.neutron [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.710436] env[69171]: DEBUG nova.scheduler.client.report [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 730.730926] env[69171]: DEBUG nova.scheduler.client.report [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 730.730926] env[69171]: DEBUG nova.compute.provider_tree [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 730.749407] env[69171]: DEBUG nova.scheduler.client.report [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 730.780584] env[69171]: DEBUG nova.scheduler.client.report [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 731.036386] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Releasing lock "refresh_cache-e1498893-a512-4b47-b22a-66c9e78df3de" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.038557] env[69171]: DEBUG nova.compute.manager [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 731.038557] env[69171]: DEBUG nova.compute.manager [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 731.038557] env[69171]: DEBUG nova.network.neutron [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 731.088579] env[69171]: INFO nova.scheduler.client.report [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Deleted allocations for instance 9f311bf0-31a1-46a9-911d-4e393c2279be [ 731.099771] env[69171]: DEBUG nova.network.neutron [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 731.208536] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2ff3fcc-2dab-4148-945f-827400ec2785 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.221805] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99837738-0c90-4030-90e5-b54c0e552a10 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.262605] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4875cb-2ddd-4e02-a170-4623046c530c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.273251] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e065be31-69d3-43d6-a084-5c715d2da2b6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.293305] env[69171]: DEBUG nova.compute.provider_tree [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 731.566063] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Acquiring lock "14990d60-4a6a-4896-816f-7bd7d5af19a5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.566063] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Lock "14990d60-4a6a-4896-816f-7bd7d5af19a5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.613144] env[69171]: DEBUG nova.network.neutron [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 731.613144] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f44a734f-ca63-4777-97ae-06ba108d0d76 tempest-ImagesOneServerNegativeTestJSON-412142896 tempest-ImagesOneServerNegativeTestJSON-412142896-project-member] Lock "9f311bf0-31a1-46a9-911d-4e393c2279be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.793s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.747752] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Acquiring lock "17c421bd-a02f-41a6-9074-8c143e423c79" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.747752] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Lock "17c421bd-a02f-41a6-9074-8c143e423c79" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.821299] env[69171]: ERROR nova.scheduler.client.report [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [req-76a5b942-2e4a-4658-9e7d-a0728c707a46] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-76a5b942-2e4a-4658-9e7d-a0728c707a46"}]} [ 731.822011] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.155s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 731.822703] env[69171]: ERROR nova.compute.manager [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 731.822703] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Traceback (most recent call last): [ 731.822703] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 731.822703] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] yield [ 731.822703] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 731.822703] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] self.set_inventory_for_provider( [ 731.822703] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 731.822703] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-76a5b942-2e4a-4658-9e7d-a0728c707a46"}]} [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] During handling of the above exception, another exception occurred: [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Traceback (most recent call last): [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] with self.rt.instance_claim(context, instance, node, allocs, [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 731.823236] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] return f(*args, **kwargs) [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] self._update(elevated, cn) [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] self._update_to_placement(context, compute_node, startup) [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] return attempt.get(self._wrap_exception) [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] six.reraise(self.value[0], self.value[1], self.value[2]) [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] raise value [ 731.823623] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] self.reportclient.update_from_provider_tree( [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] with catch_all(pd.uuid): [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] self.gen.throw(typ, value, traceback) [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] raise exception.ResourceProviderSyncFailed() [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 731.823990] env[69171]: ERROR nova.compute.manager [instance: d615fad6-4666-4a8d-a580-fc501abe501a] [ 731.824574] env[69171]: DEBUG nova.compute.utils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 731.825897] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.316s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 731.827520] env[69171]: INFO nova.compute.claims [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.830603] env[69171]: DEBUG nova.compute.manager [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Build of instance d615fad6-4666-4a8d-a580-fc501abe501a was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 731.831284] env[69171]: DEBUG nova.compute.manager [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 731.831284] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "refresh_cache-d615fad6-4666-4a8d-a580-fc501abe501a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.831284] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquired lock "refresh_cache-d615fad6-4666-4a8d-a580-fc501abe501a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.831526] env[69171]: DEBUG nova.network.neutron [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 732.118924] env[69171]: DEBUG nova.compute.manager [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 732.123598] env[69171]: INFO nova.compute.manager [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] [instance: e1498893-a512-4b47-b22a-66c9e78df3de] Took 1.08 seconds to deallocate network for instance. [ 732.547653] env[69171]: DEBUG nova.network.neutron [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 732.666116] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 732.786323] env[69171]: DEBUG nova.network.neutron [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.866514] env[69171]: DEBUG nova.scheduler.client.report [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 732.882106] env[69171]: DEBUG nova.scheduler.client.report [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 732.882341] env[69171]: DEBUG nova.compute.provider_tree [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 732.896875] env[69171]: DEBUG nova.scheduler.client.report [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 732.922131] env[69171]: DEBUG nova.scheduler.client.report [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 733.166727] env[69171]: INFO nova.scheduler.client.report [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Deleted allocations for instance e1498893-a512-4b47-b22a-66c9e78df3de [ 733.293022] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Releasing lock "refresh_cache-d615fad6-4666-4a8d-a580-fc501abe501a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.293022] env[69171]: DEBUG nova.compute.manager [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 733.293022] env[69171]: DEBUG nova.compute.manager [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 733.293022] env[69171]: DEBUG nova.network.neutron [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 733.324273] env[69171]: DEBUG nova.network.neutron [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 733.326451] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f295194-9363-454f-936f-bb1059af1b89 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.340273] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ccac07-4a8f-40e7-a4cb-702914f85286 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.381023] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85a79d4b-1c75-41c0-a4a7-80adf6b01d3a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.387416] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-075dab92-1690-4005-bbdf-bcd1a372e551 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.401955] env[69171]: DEBUG nova.compute.provider_tree [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 733.681300] env[69171]: DEBUG oslo_concurrency.lockutils [None req-12b35e3e-31f6-4228-954b-d385369e8f13 tempest-ServerRescueNegativeTestJSON-443950041 tempest-ServerRescueNegativeTestJSON-443950041-project-member] Lock "e1498893-a512-4b47-b22a-66c9e78df3de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.037s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.691235] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "155b53f6-a42a-48ad-bbfb-a99fc9eba789" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.691492] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "155b53f6-a42a-48ad-bbfb-a99fc9eba789" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.832668] env[69171]: DEBUG nova.network.neutron [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.931379] env[69171]: ERROR nova.scheduler.client.report [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [req-afa6039c-7189-4aff-ac3d-1f78d02e2a84] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-afa6039c-7189-4aff-ac3d-1f78d02e2a84"}]} [ 733.931379] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.105s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.931743] env[69171]: ERROR nova.compute.manager [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 733.931743] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Traceback (most recent call last): [ 733.931743] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 733.931743] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] yield [ 733.931743] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 733.931743] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] self.set_inventory_for_provider( [ 733.931743] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 733.931743] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-afa6039c-7189-4aff-ac3d-1f78d02e2a84"}]} [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] During handling of the above exception, another exception occurred: [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Traceback (most recent call last): [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] with self.rt.instance_claim(context, instance, node, allocs, [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 733.932023] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] return f(*args, **kwargs) [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] self._update(elevated, cn) [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] self._update_to_placement(context, compute_node, startup) [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] return attempt.get(self._wrap_exception) [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] six.reraise(self.value[0], self.value[1], self.value[2]) [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] raise value [ 733.932346] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] self.reportclient.update_from_provider_tree( [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] with catch_all(pd.uuid): [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] self.gen.throw(typ, value, traceback) [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] raise exception.ResourceProviderSyncFailed() [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 733.932774] env[69171]: ERROR nova.compute.manager [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] [ 733.933122] env[69171]: DEBUG nova.compute.utils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 733.935171] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.906s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.935171] env[69171]: DEBUG oslo_concurrency.lockutils [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.935171] env[69171]: INFO nova.compute.manager [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] [instance: 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03] Successfully reverted task state from None on failure for instance. [ 733.937592] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.448s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.937789] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.937958] env[69171]: INFO nova.compute.manager [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Successfully reverted task state from None on failure for instance. [ 733.942439] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.405s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.944817] env[69171]: INFO nova.compute.claims [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server [None req-11547bb5-9608-464b-9f72-ea1b4db6e816 tempest-ServerDiagnosticsTest-22928243 tempest-ServerDiagnosticsTest-22928243-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server yield [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-011aea72-c900-4f8f-96a3-5563c122a635"}]} [ 733.949280] env[69171]: ERROR oslo_messaging.rpc.server [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 733.949770] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 733.950343] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3354, in terminate_instance [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3349, in do_terminate_instance [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 733.950946] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server raise value [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 733.951673] env[69171]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 733.952229] env[69171]: ERROR oslo_messaging.rpc.server [ 733.952229] env[69171]: DEBUG nova.compute.manager [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Build of instance f8c018fc-4d2e-4d71-ba99-7ca512c122df was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 733.952229] env[69171]: DEBUG nova.compute.manager [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 733.952572] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Acquiring lock "refresh_cache-f8c018fc-4d2e-4d71-ba99-7ca512c122df" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 733.952572] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Acquired lock "refresh_cache-f8c018fc-4d2e-4d71-ba99-7ca512c122df" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 733.952572] env[69171]: DEBUG nova.network.neutron [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server [None req-6f74da8a-bbd6-47b7-8f47-b8739416a035 tempest-DeleteServersAdminTestJSON-1164357483 tempest-DeleteServersAdminTestJSON-1164357483-project-admin] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server yield [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-73bdc089-765c-463f-8276-fb8c03a04fba"}]} [ 733.963672] env[69171]: ERROR oslo_messaging.rpc.server [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 733.964068] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 733.964609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3354, in terminate_instance [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3349, in do_terminate_instance [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 733.965537] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server raise value [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 733.966094] env[69171]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 733.966797] env[69171]: ERROR oslo_messaging.rpc.server [ 734.184830] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 734.334947] env[69171]: INFO nova.compute.manager [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: d615fad6-4666-4a8d-a580-fc501abe501a] Took 1.04 seconds to deallocate network for instance. [ 734.496960] env[69171]: DEBUG nova.network.neutron [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 734.656556] env[69171]: DEBUG nova.network.neutron [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.716060] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 734.985079] env[69171]: DEBUG nova.scheduler.client.report [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 735.002131] env[69171]: DEBUG nova.scheduler.client.report [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 735.002382] env[69171]: DEBUG nova.compute.provider_tree [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 735.017748] env[69171]: DEBUG nova.scheduler.client.report [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 735.039783] env[69171]: DEBUG nova.scheduler.client.report [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 735.158656] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Releasing lock "refresh_cache-f8c018fc-4d2e-4d71-ba99-7ca512c122df" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 735.158904] env[69171]: DEBUG nova.compute.manager [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 735.159106] env[69171]: DEBUG nova.compute.manager [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] [instance: f8c018fc-4d2e-4d71-ba99-7ca512c122df] Skipping network deallocation for instance since networking was not requested. {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2286}} [ 735.387945] env[69171]: INFO nova.scheduler.client.report [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Deleted allocations for instance d615fad6-4666-4a8d-a580-fc501abe501a [ 735.403646] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquiring lock "ea94a5c3-226a-4866-8d96-2e5c341d4636" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.405267] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "ea94a5c3-226a-4866-8d96-2e5c341d4636" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.442875] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a62b61f8-2520-417f-8dcf-28c0580c09e7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.451382] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628a3399-d4dd-4106-829e-0a095fe63edd {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.483529] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c22ab965-0fd9-493b-9bf0-ac7e5168be79 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.491525] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab44b3b5-bfc5-4a25-a91a-a991fc039333 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.506203] env[69171]: DEBUG nova.compute.provider_tree [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 735.907402] env[69171]: DEBUG oslo_concurrency.lockutils [None req-91c40c51-79d5-4f1b-a4ad-644c48f0eb55 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "d615fad6-4666-4a8d-a580-fc501abe501a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.011s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.043571] env[69171]: ERROR nova.scheduler.client.report [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [req-156f421c-1bfd-4899-aa96-2a40b0334567] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-156f421c-1bfd-4899-aa96-2a40b0334567"}]} [ 736.043943] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.102s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.044710] env[69171]: ERROR nova.compute.manager [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 736.044710] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Traceback (most recent call last): [ 736.044710] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 736.044710] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] yield [ 736.044710] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 736.044710] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] self.set_inventory_for_provider( [ 736.044710] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 736.044710] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-156f421c-1bfd-4899-aa96-2a40b0334567"}]} [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] During handling of the above exception, another exception occurred: [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Traceback (most recent call last): [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] with self.rt.instance_claim(context, instance, node, allocs, [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 736.045082] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] return f(*args, **kwargs) [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] self._update(elevated, cn) [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] self._update_to_placement(context, compute_node, startup) [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] return attempt.get(self._wrap_exception) [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] six.reraise(self.value[0], self.value[1], self.value[2]) [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] raise value [ 736.045433] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] self.reportclient.update_from_provider_tree( [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] with catch_all(pd.uuid): [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] self.gen.throw(typ, value, traceback) [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] raise exception.ResourceProviderSyncFailed() [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 736.045842] env[69171]: ERROR nova.compute.manager [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] [ 736.046248] env[69171]: DEBUG nova.compute.utils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.047189] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.911s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.047189] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.047189] env[69171]: INFO nova.compute.manager [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] [instance: 6a29d495-c938-41ad-97b6-1331b6eff589] Successfully reverted task state from None on failure for instance. [ 736.049238] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.904s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.051885] env[69171]: INFO nova.compute.claims [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 736.054678] env[69171]: DEBUG nova.compute.manager [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Build of instance 4c596d23-7da7-4367-a60c-fe3b789516d9 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 736.055129] env[69171]: DEBUG nova.compute.manager [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 736.055356] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Acquiring lock "refresh_cache-4c596d23-7da7-4367-a60c-fe3b789516d9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.055501] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Acquired lock "refresh_cache-4c596d23-7da7-4367-a60c-fe3b789516d9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.055657] env[69171]: DEBUG nova.network.neutron [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server [None req-5b260f01-fb60-4512-af03-e2a1296ef6fc tempest-ServerDiagnosticsV248Test-1946156274 tempest-ServerDiagnosticsV248Test-1946156274-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server yield [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-86f54cf9-68a7-4b09-89e2-3019b6f1f20f"}]} [ 736.057596] env[69171]: ERROR oslo_messaging.rpc.server [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 736.057960] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 736.058496] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3354, in terminate_instance [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3349, in do_terminate_instance [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 736.059178] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server raise value [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 736.060128] env[69171]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 736.060936] env[69171]: ERROR oslo_messaging.rpc.server [ 736.243619] env[69171]: INFO nova.scheduler.client.report [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Deleted allocations for instance f8c018fc-4d2e-4d71-ba99-7ca512c122df [ 736.410982] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 736.590086] env[69171]: DEBUG nova.network.neutron [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 736.741874] env[69171]: DEBUG nova.network.neutron [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.755721] env[69171]: DEBUG oslo_concurrency.lockutils [None req-93ffd9ee-ece0-410b-bd9b-9509d2ad4a03 tempest-ServersAaction247Test-2092532634 tempest-ServersAaction247Test-2092532634-project-member] Lock "f8c018fc-4d2e-4d71-ba99-7ca512c122df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.968s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.935923] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.089223] env[69171]: DEBUG nova.scheduler.client.report [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 737.111345] env[69171]: DEBUG nova.scheduler.client.report [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 737.113045] env[69171]: DEBUG nova.compute.provider_tree [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 737.135388] env[69171]: DEBUG nova.scheduler.client.report [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 737.163374] env[69171]: DEBUG nova.scheduler.client.report [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 737.246241] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Releasing lock "refresh_cache-4c596d23-7da7-4367-a60c-fe3b789516d9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.246482] env[69171]: DEBUG nova.compute.manager [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 737.246671] env[69171]: DEBUG nova.compute.manager [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] [instance: 4c596d23-7da7-4367-a60c-fe3b789516d9] Skipping network deallocation for instance since networking was not requested. {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2286}} [ 737.260338] env[69171]: DEBUG nova.compute.manager [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 737.570971] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a715bc5-4aca-41c3-9f38-3a052fd73595 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.583966] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92d14032-ea80-4b3b-a10f-daa65d04ef11 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.627495] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e754083-2f1c-4640-804d-46ae227c6ba8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.641215] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb96f0d1-fad1-4886-9a6e-4f3de27db26f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.660400] env[69171]: DEBUG nova.compute.provider_tree [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 737.796215] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.194166] env[69171]: ERROR nova.scheduler.client.report [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [req-2c8da8d2-ba72-481b-9692-88af17a5e21b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2c8da8d2-ba72-481b-9692-88af17a5e21b"}]} [ 738.194166] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.145s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.195149] env[69171]: ERROR nova.compute.manager [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 738.195149] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Traceback (most recent call last): [ 738.195149] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 738.195149] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] yield [ 738.195149] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 738.195149] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] self.set_inventory_for_provider( [ 738.195149] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 738.195149] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2c8da8d2-ba72-481b-9692-88af17a5e21b"}]} [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] During handling of the above exception, another exception occurred: [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Traceback (most recent call last): [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] with self.rt.instance_claim(context, instance, node, allocs, [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 738.195463] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] return f(*args, **kwargs) [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] self._update(elevated, cn) [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] self._update_to_placement(context, compute_node, startup) [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] return attempt.get(self._wrap_exception) [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] six.reraise(self.value[0], self.value[1], self.value[2]) [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] raise value [ 738.195762] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] self.reportclient.update_from_provider_tree( [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] with catch_all(pd.uuid): [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] self.gen.throw(typ, value, traceback) [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] raise exception.ResourceProviderSyncFailed() [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 738.196204] env[69171]: ERROR nova.compute.manager [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] [ 738.196553] env[69171]: DEBUG nova.compute.utils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.197897] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.928s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.201838] env[69171]: DEBUG nova.compute.manager [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Build of instance 62f1950f-631d-4f07-8526-2155d4e1bc7c was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 738.202265] env[69171]: DEBUG nova.compute.manager [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 738.202527] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "refresh_cache-62f1950f-631d-4f07-8526-2155d4e1bc7c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.202637] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquired lock "refresh_cache-62f1950f-631d-4f07-8526-2155d4e1bc7c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.204031] env[69171]: DEBUG nova.network.neutron [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 738.222052] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 738.243871] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 738.244187] env[69171]: DEBUG nova.compute.provider_tree [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 738.263432] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 738.287248] env[69171]: DEBUG nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 738.295357] env[69171]: INFO nova.scheduler.client.report [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Deleted allocations for instance 4c596d23-7da7-4367-a60c-fe3b789516d9 [ 738.572563] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f3d62fa-233e-4b07-9d08-11fe8197291b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.579717] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5a24bae-1ae6-4a40-b37c-536636fb9959 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.617428] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-155d1a09-cd17-4248-8d17-21ff2ba0a837 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.626498] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e211688-fe72-40fd-80d9-e5fcd2d312a4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.643957] env[69171]: DEBUG nova.compute.provider_tree [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 738.733325] env[69171]: DEBUG nova.network.neutron [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 738.811017] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6acb8227-293b-41df-8d1a-9f2ca51b6451 tempest-ServersListShow296Test-1431838549 tempest-ServersListShow296Test-1431838549-project-member] Lock "4c596d23-7da7-4367-a60c-fe3b789516d9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.304s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.861642] env[69171]: DEBUG nova.network.neutron [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.127738] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquiring lock "d0facc29-338c-4d0d-a9f3-a37781c19b68" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.128029] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "d0facc29-338c-4d0d-a9f3-a37781c19b68" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.176906] env[69171]: ERROR nova.scheduler.client.report [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [req-939650d8-6b58-4831-bd48-1d917ee4a1fd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-939650d8-6b58-4831-bd48-1d917ee4a1fd"}]}: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 739.176906] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.978s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 739.177047] env[69171]: WARNING nova.compute.manager [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Failed to revert task state for instance. Error: Failed to synchronize the placement service with resource provider information supplied by the compute host.: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 739.188523] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.891s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.190361] env[69171]: INFO nova.compute.claims [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server [None req-c03bc1f5-a8f2-4fe3-bc1b-eea79b714f4c tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server yield [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-198b453c-bcf0-4250-b85a-4091d3202998"}]} [ 739.194119] env[69171]: ERROR oslo_messaging.rpc.server [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 739.194609] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 739.194972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5767, in prep_resize [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server self._reschedule_resize_or_reraise(context, instance, [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5847, in _reschedule_resize_or_reraise [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server raise exc [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5749, in prep_resize [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server self._prep_resize(context, image, instance, [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 5662, in _prep_resize [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server with self.rt.resize_claim( [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 238, in resize_claim [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server return self._move_claim( [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 378, in _move_claim [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server self._update(elevated, cn) [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 739.195315] env[69171]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server raise value [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 739.195675] env[69171]: ERROR oslo_messaging.rpc.server [ 739.316852] env[69171]: DEBUG nova.compute.manager [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 739.338018] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Releasing lock "refresh_cache-62f1950f-631d-4f07-8526-2155d4e1bc7c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.338018] env[69171]: DEBUG nova.compute.manager [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 739.338018] env[69171]: DEBUG nova.compute.manager [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 739.338018] env[69171]: DEBUG nova.network.neutron [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 739.445075] env[69171]: DEBUG nova.network.neutron [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 739.846651] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.947324] env[69171]: DEBUG nova.network.neutron [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.228992] env[69171]: DEBUG nova.scheduler.client.report [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 740.249191] env[69171]: DEBUG nova.scheduler.client.report [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 740.249558] env[69171]: DEBUG nova.compute.provider_tree [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 740.267455] env[69171]: DEBUG nova.scheduler.client.report [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 740.293455] env[69171]: DEBUG nova.scheduler.client.report [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 740.455106] env[69171]: INFO nova.compute.manager [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: 62f1950f-631d-4f07-8526-2155d4e1bc7c] Took 1.12 seconds to deallocate network for instance. [ 740.697398] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4041dec9-80d1-4095-9c9a-aa733126d82a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.709781] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e6f8b44-526d-4416-a96f-e549facaa517 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.748853] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-592892db-72af-4f7c-ba52-088f8124dddd {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.757542] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ece3c69-083d-4422-a64a-8bfdb1eb5283 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.775040] env[69171]: DEBUG nova.compute.provider_tree [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 740.800908] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "8e6a9d73-6705-484c-82e0-70c6a6860211" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.801187] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.801392] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "8e6a9d73-6705-484c-82e0-70c6a6860211-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.801577] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.801842] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.804863] env[69171]: INFO nova.compute.manager [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Terminating instance [ 740.813165] env[69171]: DEBUG nova.compute.manager [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Start destroying the instance on the hypervisor. {{(pid=69171) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3151}} [ 740.813165] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Destroying instance {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 740.813165] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3728f425-95d9-47b9-8c5c-2b1f060f6cd9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.819898] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Powering off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 740.820173] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6cf197a7-1d43-4d74-a413-bb80df95483b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.834027] env[69171]: DEBUG oslo_vmware.api [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 740.834027] env[69171]: value = "task-4235929" [ 740.834027] env[69171]: _type = "Task" [ 740.834027] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.846804] env[69171]: DEBUG oslo_vmware.api [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235929, 'name': PowerOffVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.302507] env[69171]: ERROR nova.scheduler.client.report [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [req-c268562c-6861-4b2b-834c-fedd7acba5ae] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c268562c-6861-4b2b-834c-fedd7acba5ae"}]} [ 741.302811] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.114s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.303641] env[69171]: ERROR nova.compute.manager [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 741.303641] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Traceback (most recent call last): [ 741.303641] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 741.303641] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] yield [ 741.303641] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 741.303641] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] self.set_inventory_for_provider( [ 741.303641] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 741.303641] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c268562c-6861-4b2b-834c-fedd7acba5ae"}]} [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] During handling of the above exception, another exception occurred: [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Traceback (most recent call last): [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] with self.rt.instance_claim(context, instance, node, allocs, [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 741.303940] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] return f(*args, **kwargs) [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] self._update(elevated, cn) [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] self._update_to_placement(context, compute_node, startup) [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] return attempt.get(self._wrap_exception) [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] six.reraise(self.value[0], self.value[1], self.value[2]) [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] raise value [ 741.304443] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] self.reportclient.update_from_provider_tree( [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] with catch_all(pd.uuid): [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] self.gen.throw(typ, value, traceback) [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] raise exception.ResourceProviderSyncFailed() [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 741.304771] env[69171]: ERROR nova.compute.manager [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] [ 741.305416] env[69171]: DEBUG nova.compute.utils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 741.307511] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.448s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.309654] env[69171]: INFO nova.compute.claims [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 741.312309] env[69171]: DEBUG nova.compute.manager [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Build of instance bd91753e-aced-4c58-a6f1-f1baabbdd8c4 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 741.312732] env[69171]: DEBUG nova.compute.manager [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 741.312959] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "refresh_cache-bd91753e-aced-4c58-a6f1-f1baabbdd8c4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 741.313225] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquired lock "refresh_cache-bd91753e-aced-4c58-a6f1-f1baabbdd8c4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 741.313421] env[69171]: DEBUG nova.network.neutron [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 741.345767] env[69171]: DEBUG oslo_vmware.api [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235929, 'name': PowerOffVM_Task, 'duration_secs': 0.226519} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.346226] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Powered off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1507}} [ 741.346952] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Unregistering the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1094}} [ 741.347246] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-88e937ed-100b-4b93-bc81-22373c8fd8e3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.412291] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Unregistered the VM {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1097}} [ 741.413376] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Deleting contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1107}} [ 741.413376] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Deleting the datastore file [datastore1] 8e6a9d73-6705-484c-82e0-70c6a6860211 {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 741.413376] env[69171]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b41e5652-ae6d-4db7-871a-017ea49a74fb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.421381] env[69171]: DEBUG oslo_vmware.api [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for the task: (returnval){ [ 741.421381] env[69171]: value = "task-4235931" [ 741.421381] env[69171]: _type = "Task" [ 741.421381] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.430187] env[69171]: DEBUG oslo_vmware.api [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235931, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.493999] env[69171]: INFO nova.scheduler.client.report [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Deleted allocations for instance 62f1950f-631d-4f07-8526-2155d4e1bc7c [ 741.864227] env[69171]: DEBUG nova.network.neutron [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 741.943024] env[69171]: DEBUG oslo_vmware.api [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Task: {'id': task-4235931, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.200244} completed successfully. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.943024] env[69171]: DEBUG nova.virt.vmwareapi.ds_util [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Deleted the datastore file {{(pid=69171) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 741.943024] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Deleted contents of the VM from datastore datastore1 {{(pid=69171) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1117}} [ 741.943024] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Instance destroyed {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 741.943024] env[69171]: INFO nova.compute.manager [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Took 1.13 seconds to destroy the instance on the hypervisor. [ 741.943328] env[69171]: DEBUG oslo.service.loopingcall [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 741.943328] env[69171]: DEBUG nova.compute.manager [-] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 741.943328] env[69171]: DEBUG nova.network.neutron [-] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.005098] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d874429-3aa3-49fa-bb09-c697137c1235 tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "62f1950f-631d-4f07-8526-2155d4e1bc7c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.909s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.167783] env[69171]: DEBUG nova.network.neutron [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 742.277255] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7b994a0a-dc43-4089-b160-f31cc15302f9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "87075f00-32ea-4785-ace5-e88d07fe351c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.277588] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7b994a0a-dc43-4089-b160-f31cc15302f9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "87075f00-32ea-4785-ace5-e88d07fe351c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.353224] env[69171]: DEBUG nova.scheduler.client.report [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 742.368507] env[69171]: DEBUG nova.scheduler.client.report [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 742.368707] env[69171]: DEBUG nova.compute.provider_tree [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 742.381894] env[69171]: DEBUG nova.scheduler.client.report [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 742.400173] env[69171]: DEBUG nova.scheduler.client.report [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 742.508724] env[69171]: DEBUG nova.compute.manager [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 742.670886] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Releasing lock "refresh_cache-bd91753e-aced-4c58-a6f1-f1baabbdd8c4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.671443] env[69171]: DEBUG nova.compute.manager [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 742.671780] env[69171]: DEBUG nova.compute.manager [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 742.672569] env[69171]: DEBUG nova.network.neutron [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 742.721622] env[69171]: DEBUG nova.network.neutron [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 742.738031] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd43736-2f97-4c4a-9522-d4001060ea7e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.746430] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d21c3d4-df07-4646-87ff-c2300634bcc5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.782839] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c03ba5c-dfdc-47d4-a4e5-b9b86ea645e0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.795775] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b43f2e8-934c-45a7-8f18-53d40378846e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.815236] env[69171]: DEBUG nova.compute.provider_tree [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 743.037837] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.224688] env[69171]: DEBUG nova.network.neutron [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.351254] env[69171]: ERROR nova.scheduler.client.report [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [req-ba16e44d-332d-4649-8276-c02e6ae83986] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ba16e44d-332d-4649-8276-c02e6ae83986"}]} [ 743.351745] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.044s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.352396] env[69171]: ERROR nova.compute.manager [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 743.352396] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Traceback (most recent call last): [ 743.352396] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 743.352396] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] yield [ 743.352396] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 743.352396] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] self.set_inventory_for_provider( [ 743.352396] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 743.352396] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ba16e44d-332d-4649-8276-c02e6ae83986"}]} [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] During handling of the above exception, another exception occurred: [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Traceback (most recent call last): [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] with self.rt.instance_claim(context, instance, node, allocs, [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 743.352629] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] return f(*args, **kwargs) [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] self._update(elevated, cn) [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] self._update_to_placement(context, compute_node, startup) [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] return attempt.get(self._wrap_exception) [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] six.reraise(self.value[0], self.value[1], self.value[2]) [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] raise value [ 743.352874] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] self.reportclient.update_from_provider_tree( [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] with catch_all(pd.uuid): [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] self.gen.throw(typ, value, traceback) [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] raise exception.ResourceProviderSyncFailed() [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 743.353255] env[69171]: ERROR nova.compute.manager [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] [ 743.354730] env[69171]: DEBUG nova.compute.utils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.356325] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.769s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.358698] env[69171]: INFO nova.compute.claims [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 743.363218] env[69171]: DEBUG nova.compute.manager [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Build of instance f1a31770-1165-4143-be50-0e6a6ce2f041 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 743.363819] env[69171]: DEBUG nova.compute.manager [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 743.365189] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Acquiring lock "refresh_cache-f1a31770-1165-4143-be50-0e6a6ce2f041" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.365423] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Acquired lock "refresh_cache-f1a31770-1165-4143-be50-0e6a6ce2f041" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.365657] env[69171]: DEBUG nova.network.neutron [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 743.463248] env[69171]: DEBUG nova.network.neutron [-] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.733749] env[69171]: INFO nova.compute.manager [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: bd91753e-aced-4c58-a6f1-f1baabbdd8c4] Took 1.06 seconds to deallocate network for instance. [ 743.862502] env[69171]: DEBUG nova.compute.manager [req-06b8a2e5-7e9b-471a-85c4-535f326ade5b req-cb4ad2e6-06aa-4b06-9e12-eab2d18c4d7d service nova] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Received event network-vif-deleted-547da9e0-630e-4393-95ef-f0b0086a5b94 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 743.910671] env[69171]: DEBUG nova.network.neutron [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 743.967947] env[69171]: INFO nova.compute.manager [-] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Took 2.03 seconds to deallocate network for instance. [ 744.100943] env[69171]: DEBUG nova.network.neutron [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.238176] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Acquiring lock "4c5586e7-333c-44b0-aef9-4d1b7988826d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.238176] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Lock "4c5586e7-333c-44b0-aef9-4d1b7988826d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 744.398293] env[69171]: DEBUG nova.scheduler.client.report [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 744.428241] env[69171]: DEBUG nova.scheduler.client.report [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 744.428470] env[69171]: DEBUG nova.compute.provider_tree [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 744.442640] env[69171]: DEBUG nova.scheduler.client.report [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 744.475312] env[69171]: DEBUG nova.scheduler.client.report [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 744.479718] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 744.606517] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Releasing lock "refresh_cache-f1a31770-1165-4143-be50-0e6a6ce2f041" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.606788] env[69171]: DEBUG nova.compute.manager [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 744.606981] env[69171]: DEBUG nova.compute.manager [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 744.607247] env[69171]: DEBUG nova.network.neutron [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 744.660972] env[69171]: DEBUG nova.network.neutron [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 744.787131] env[69171]: INFO nova.scheduler.client.report [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Deleted allocations for instance bd91753e-aced-4c58-a6f1-f1baabbdd8c4 [ 744.889914] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177294a1-a5a2-4b23-83ed-226cb0fa0cd3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.903419] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-804d762f-701d-4187-8d8c-0c9516bff35b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.941448] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26aae6f-ae2c-4339-abbd-91c29da438e1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.951167] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ac1ce6f-a9ee-4c64-97a2-d194ae814251 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.967034] env[69171]: DEBUG nova.compute.provider_tree [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 745.170570] env[69171]: DEBUG nova.network.neutron [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.260289] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.260868] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.261132] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.261288] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.261455] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.265035] env[69171]: INFO nova.compute.manager [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Terminating instance [ 745.269613] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.269613] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquired lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.269613] env[69171]: DEBUG nova.network.neutron [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.300209] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1199304b-ad7d-4d2c-ad36-033649529c8f tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "bd91753e-aced-4c58-a6f1-f1baabbdd8c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.041s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.498619] env[69171]: ERROR nova.scheduler.client.report [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [req-863642ca-db2a-4f04-b34b-18569b740415] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-863642ca-db2a-4f04-b34b-18569b740415"}]} [ 745.498619] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.141s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.498904] env[69171]: ERROR nova.compute.manager [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 745.498904] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Traceback (most recent call last): [ 745.498904] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 745.498904] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] yield [ 745.498904] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 745.498904] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] self.set_inventory_for_provider( [ 745.498904] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 745.498904] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-863642ca-db2a-4f04-b34b-18569b740415"}]} [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] During handling of the above exception, another exception occurred: [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Traceback (most recent call last): [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] with self.rt.instance_claim(context, instance, node, allocs, [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 745.499104] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] return f(*args, **kwargs) [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] self._update(elevated, cn) [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] self._update_to_placement(context, compute_node, startup) [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] return attempt.get(self._wrap_exception) [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] six.reraise(self.value[0], self.value[1], self.value[2]) [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] raise value [ 745.499322] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] self.reportclient.update_from_provider_tree( [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] with catch_all(pd.uuid): [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] self.gen.throw(typ, value, traceback) [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] raise exception.ResourceProviderSyncFailed() [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 745.499586] env[69171]: ERROR nova.compute.manager [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] [ 745.499811] env[69171]: DEBUG nova.compute.utils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.504430] env[69171]: DEBUG nova.compute.manager [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Build of instance 3b8a2f7d-51a2-42f1-ad30-12e3a1083037 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 745.504430] env[69171]: DEBUG nova.compute.manager [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 745.504430] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquiring lock "refresh_cache-3b8a2f7d-51a2-42f1-ad30-12e3a1083037" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.505253] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquired lock "refresh_cache-3b8a2f7d-51a2-42f1-ad30-12e3a1083037" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.505253] env[69171]: DEBUG nova.network.neutron [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 745.506982] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.010s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.509767] env[69171]: INFO nova.compute.claims [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 745.678853] env[69171]: INFO nova.compute.manager [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] [instance: f1a31770-1165-4143-be50-0e6a6ce2f041] Took 1.07 seconds to deallocate network for instance. [ 745.749528] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "6cdbe208-0eb2-4493-a477-daeb70739f95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.749785] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "6cdbe208-0eb2-4493-a477-daeb70739f95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.807031] env[69171]: DEBUG nova.compute.manager [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 745.818947] env[69171]: DEBUG nova.network.neutron [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 745.975952] env[69171]: DEBUG nova.network.neutron [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.071971] env[69171]: DEBUG nova.network.neutron [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.335383] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 746.357258] env[69171]: DEBUG nova.network.neutron [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.479545] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Releasing lock "refresh_cache-fa5997fa-817c-42b6-a7fa-b3b85a109b6d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.480630] env[69171]: DEBUG nova.compute.manager [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Start destroying the instance on the hypervisor. {{(pid=69171) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3151}} [ 746.480630] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Destroying instance {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 746.480992] env[69171]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-32eb8084-9c97-4a0a-a814-06219ab08f4e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.494053] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43e96ca8-c855-41fb-b54b-5cc7cd44d0e3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.524905] env[69171]: WARNING nova.virt.vmwareapi.vmops [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fa5997fa-817c-42b6-a7fa-b3b85a109b6d could not be found. [ 746.525175] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance destroyed {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 746.525295] env[69171]: INFO nova.compute.manager [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 746.525580] env[69171]: DEBUG oslo.service.loopingcall [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.526317] env[69171]: DEBUG nova.compute.manager [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 746.526317] env[69171]: DEBUG nova.network.neutron [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 746.551690] env[69171]: DEBUG nova.scheduler.client.report [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 746.564121] env[69171]: DEBUG nova.network.neutron [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.573703] env[69171]: DEBUG nova.scheduler.client.report [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 746.576173] env[69171]: DEBUG nova.compute.provider_tree [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 746.589154] env[69171]: DEBUG nova.scheduler.client.report [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 746.615825] env[69171]: DEBUG nova.scheduler.client.report [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 746.724372] env[69171]: INFO nova.scheduler.client.report [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Deleted allocations for instance f1a31770-1165-4143-be50-0e6a6ce2f041 [ 746.861277] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Releasing lock "refresh_cache-3b8a2f7d-51a2-42f1-ad30-12e3a1083037" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.861892] env[69171]: DEBUG nova.compute.manager [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 746.861892] env[69171]: DEBUG nova.compute.manager [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 746.861998] env[69171]: DEBUG nova.network.neutron [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 746.898888] env[69171]: DEBUG nova.network.neutron [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 746.937068] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 746.937575] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.004991] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316335d1-f368-4fe6-be23-9cf9c6250e45 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.014840] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998184f2-c471-4d82-8a21-bc02c9595689 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.052335] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91c74ddb-43d1-49ef-b4b5-5beb48cec532 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.062066] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86dbe5da-e4ab-4f2a-afd0-5af8603dddb4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.075726] env[69171]: DEBUG nova.network.neutron [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.077293] env[69171]: DEBUG nova.compute.provider_tree [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 747.232294] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba1c0e1c-7999-4b41-a7f1-b34d044c30cb tempest-FloatingIPsAssociationNegativeTestJSON-1901514175 tempest-FloatingIPsAssociationNegativeTestJSON-1901514175-project-member] Lock "f1a31770-1165-4143-be50-0e6a6ce2f041" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.457s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.404266] env[69171]: DEBUG nova.network.neutron [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.444323] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 747.444545] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Starting heal instance info cache {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9929}} [ 747.444696] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Rebuilding the list of instances to heal {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9933}} [ 747.582395] env[69171]: INFO nova.compute.manager [-] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Took 1.06 seconds to deallocate network for instance. [ 747.612166] env[69171]: ERROR nova.scheduler.client.report [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [req-419e5238-5996-49a2-b0b0-c1eed927089c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-419e5238-5996-49a2-b0b0-c1eed927089c"}]} [ 747.613492] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.106s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.613492] env[69171]: ERROR nova.compute.manager [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 747.613492] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Traceback (most recent call last): [ 747.613492] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 747.613492] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] yield [ 747.613492] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 747.613492] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] self.set_inventory_for_provider( [ 747.613492] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 747.613492] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-419e5238-5996-49a2-b0b0-c1eed927089c"}]} [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] During handling of the above exception, another exception occurred: [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Traceback (most recent call last): [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] with self.rt.instance_claim(context, instance, node, allocs, [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 747.614448] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] return f(*args, **kwargs) [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] self._update(elevated, cn) [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] self._update_to_placement(context, compute_node, startup) [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] return attempt.get(self._wrap_exception) [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] six.reraise(self.value[0], self.value[1], self.value[2]) [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] raise value [ 747.614851] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] self.reportclient.update_from_provider_tree( [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] with catch_all(pd.uuid): [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] self.gen.throw(typ, value, traceback) [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] raise exception.ResourceProviderSyncFailed() [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 747.615354] env[69171]: ERROR nova.compute.manager [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] [ 747.615627] env[69171]: DEBUG nova.compute.utils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.615627] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.951s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.616536] env[69171]: INFO nova.compute.claims [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.619198] env[69171]: DEBUG nova.compute.manager [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Build of instance 363084be-ae28-433b-b7e0-d0a4eccbff6a was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 747.619625] env[69171]: DEBUG nova.compute.manager [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 747.619916] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Acquiring lock "refresh_cache-363084be-ae28-433b-b7e0-d0a4eccbff6a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.619995] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Acquired lock "refresh_cache-363084be-ae28-433b-b7e0-d0a4eccbff6a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.620162] env[69171]: DEBUG nova.network.neutron [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 747.739668] env[69171]: DEBUG nova.compute.manager [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 747.907834] env[69171]: INFO nova.compute.manager [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Took 1.05 seconds to deallocate network for instance. [ 747.949675] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Skipping network cache update for instance because it is being deleted. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9946}} [ 747.949879] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 3b8a2f7d-51a2-42f1-ad30-12e3a1083037] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 747.949959] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Skipping network cache update for instance because it is Building. {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9942}} [ 747.966010] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "9bd1c816-5248-4e48-89f8-f3b999d50bff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.967259] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "9bd1c816-5248-4e48-89f8-f3b999d50bff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 748.082925] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.082925] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquired lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.082925] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Forcefully refreshing network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 748.082925] env[69171]: DEBUG nova.objects.instance [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lazy-loading 'info_cache' on Instance uuid f8b20421-15d3-46f3-a504-172d044ff4d3 {{(pid=69171) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 748.094686] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.224354] env[69171]: DEBUG nova.network.neutron [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 748.271316] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.654729] env[69171]: DEBUG nova.scheduler.client.report [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 748.677356] env[69171]: DEBUG nova.scheduler.client.report [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 748.677356] env[69171]: DEBUG nova.compute.provider_tree [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 748.691435] env[69171]: DEBUG nova.scheduler.client.report [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 748.707515] env[69171]: DEBUG nova.network.neutron [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.715354] env[69171]: DEBUG nova.scheduler.client.report [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 748.960074] env[69171]: INFO nova.scheduler.client.report [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Deleted allocations for instance 3b8a2f7d-51a2-42f1-ad30-12e3a1083037 [ 749.210323] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e9615b-d6af-4023-8db5-548b55699379 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.217477] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Releasing lock "refresh_cache-363084be-ae28-433b-b7e0-d0a4eccbff6a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.217735] env[69171]: DEBUG nova.compute.manager [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 749.217967] env[69171]: DEBUG nova.compute.manager [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] [instance: 363084be-ae28-433b-b7e0-d0a4eccbff6a] Skipping network deallocation for instance since networking was not requested. {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2286}} [ 749.230385] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bb20775-51f4-40c5-a49b-a23390f5bed6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.274083] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731a720b-d20b-46e5-9352-34317b1d3a58 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.283883] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de1f0c9a-a678-4ad4-80ed-e1793a99d7b1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.306324] env[69171]: DEBUG nova.compute.provider_tree [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 749.476270] env[69171]: DEBUG oslo_concurrency.lockutils [None req-30d82f0a-d235-4203-91a8-b4891712f189 tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "3b8a2f7d-51a2-42f1-ad30-12e3a1083037" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.897s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.842593] env[69171]: ERROR nova.scheduler.client.report [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [req-c8a38547-6f9b-44e4-9987-eef6b6a3c898] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c8a38547-6f9b-44e4-9987-eef6b6a3c898"}]} [ 749.845291] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.845291] env[69171]: ERROR nova.compute.manager [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 749.845291] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Traceback (most recent call last): [ 749.845291] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 749.845291] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] yield [ 749.845291] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 749.845291] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] self.set_inventory_for_provider( [ 749.845291] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 749.845291] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c8a38547-6f9b-44e4-9987-eef6b6a3c898"}]} [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] During handling of the above exception, another exception occurred: [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Traceback (most recent call last): [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] with self.rt.instance_claim(context, instance, node, allocs, [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 749.846372] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] return f(*args, **kwargs) [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] self._update(elevated, cn) [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] self._update_to_placement(context, compute_node, startup) [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] return attempt.get(self._wrap_exception) [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] six.reraise(self.value[0], self.value[1], self.value[2]) [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] raise value [ 749.846691] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] self.reportclient.update_from_provider_tree( [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] with catch_all(pd.uuid): [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] self.gen.throw(typ, value, traceback) [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] raise exception.ResourceProviderSyncFailed() [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 749.847022] env[69171]: ERROR nova.compute.manager [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] [ 749.847299] env[69171]: DEBUG nova.compute.utils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 749.847299] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.130s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.848876] env[69171]: INFO nova.compute.claims [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.852121] env[69171]: DEBUG nova.compute.manager [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Build of instance e0f59163-b386-45d3-87d7-1271deebe2f2 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 749.852981] env[69171]: DEBUG nova.compute.manager [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 749.853993] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Acquiring lock "refresh_cache-e0f59163-b386-45d3-87d7-1271deebe2f2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 749.853993] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Acquired lock "refresh_cache-e0f59163-b386-45d3-87d7-1271deebe2f2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 749.853993] env[69171]: DEBUG nova.network.neutron [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 749.983082] env[69171]: DEBUG nova.compute.manager [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 750.117856] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updating instance_info_cache with network_info: [{"id": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "address": "fa:16:3e:83:7a:e9", "network": {"id": "9c2dbe2a-14db-48ac-8eb3-8be341ba476e", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.249", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "1f8548ca1c9947769d22274768a59ff7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f17856cf-7248-414b-bde6-8c90cfb4c593", "external-id": "nsx-vlan-transportzone-341", "segmentation_id": 341, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8cd0dd84-ce", "ovs_interfaceid": "8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.163930] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "9a29b0d3-9de0-476b-8559-672010f02caa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.165199] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "9a29b0d3-9de0-476b-8559-672010f02caa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 750.266579] env[69171]: INFO nova.scheduler.client.report [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Deleted allocations for instance 363084be-ae28-433b-b7e0-d0a4eccbff6a [ 750.383512] env[69171]: DEBUG nova.network.neutron [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 750.516540] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.598514] env[69171]: DEBUG nova.network.neutron [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 750.625467] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Releasing lock "refresh_cache-f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 750.625467] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updated the network info_cache for instance {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10000}} [ 750.625467] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.625467] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.625467] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.625467] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.625816] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.625816] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.625816] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=69171) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10548}} [ 750.625816] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager.update_available_resource {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 750.778508] env[69171]: DEBUG oslo_concurrency.lockutils [None req-b719abc4-4982-4042-a326-752f162996b9 tempest-ServersAdmin275Test-2005677628 tempest-ServersAdmin275Test-2005677628-project-member] Lock "363084be-ae28-433b-b7e0-d0a4eccbff6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 26.035s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.882135] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 750.900404] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 750.900713] env[69171]: DEBUG nova.compute.provider_tree [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 750.915923] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 750.941527] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 751.104995] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Releasing lock "refresh_cache-e0f59163-b386-45d3-87d7-1271deebe2f2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 751.104995] env[69171]: DEBUG nova.compute.manager [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 751.105176] env[69171]: DEBUG nova.compute.manager [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 751.105302] env[69171]: DEBUG nova.network.neutron [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 751.128132] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.128753] env[69171]: DEBUG nova.network.neutron [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 751.283892] env[69171]: DEBUG nova.compute.manager [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 751.326127] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e77157f2-8845-41a5-8e5f-bcff0579e60b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.344578] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-858c9c5c-3358-4561-83aa-21ff063a71ca {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.390860] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f033bb-28c9-44d3-91a0-78f05e7435ed {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.400208] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-196a6c93-7d22-4e3c-9b73-c078b5fbb083 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.423543] env[69171]: DEBUG nova.compute.provider_tree [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 751.635609] env[69171]: DEBUG nova.network.neutron [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.823809] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.956534] env[69171]: ERROR nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [req-e3584963-6a55-4070-bf3f-52f2eb47cec7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e3584963-6a55-4070-bf3f-52f2eb47cec7"}]} [ 751.956918] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.957554] env[69171]: ERROR nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 751.957554] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Traceback (most recent call last): [ 751.957554] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 751.957554] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] yield [ 751.957554] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 751.957554] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] self.set_inventory_for_provider( [ 751.957554] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 751.957554] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e3584963-6a55-4070-bf3f-52f2eb47cec7"}]} [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] During handling of the above exception, another exception occurred: [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Traceback (most recent call last): [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] with self.rt.instance_claim(context, instance, node, allocs, [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 751.957838] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] return f(*args, **kwargs) [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] self._update(elevated, cn) [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] self._update_to_placement(context, compute_node, startup) [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] return attempt.get(self._wrap_exception) [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] six.reraise(self.value[0], self.value[1], self.value[2]) [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] raise value [ 751.958123] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] self.reportclient.update_from_provider_tree( [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] with catch_all(pd.uuid): [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] self.gen.throw(typ, value, traceback) [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] raise exception.ResourceProviderSyncFailed() [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 751.958522] env[69171]: ERROR nova.compute.manager [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] [ 751.958824] env[69171]: DEBUG nova.compute.utils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.959658] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.024s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.963449] env[69171]: INFO nova.compute.claims [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 751.967047] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Build of instance 7a5215be-525e-4400-8972-4d6a949f5c24 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 751.968236] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 751.968236] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "refresh_cache-7a5215be-525e-4400-8972-4d6a949f5c24" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.968236] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquired lock "refresh_cache-7a5215be-525e-4400-8972-4d6a949f5c24" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.971040] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 752.138703] env[69171]: INFO nova.compute.manager [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] [instance: e0f59163-b386-45d3-87d7-1271deebe2f2] Took 1.03 seconds to deallocate network for instance. [ 752.542795] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 752.858517] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.009395] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 753.035979] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 753.036250] env[69171]: DEBUG nova.compute.provider_tree [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 753.057292] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 753.085748] env[69171]: DEBUG nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 753.202758] env[69171]: INFO nova.scheduler.client.report [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Deleted allocations for instance e0f59163-b386-45d3-87d7-1271deebe2f2 [ 753.364081] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Releasing lock "refresh_cache-7a5215be-525e-4400-8972-4d6a949f5c24" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.364081] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 753.364081] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 753.364081] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 753.399307] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 753.544448] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e93994a-cc3e-4991-90a3-042d79875bfa {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.554074] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1501f4cf-1d67-44b7-86ed-9643c50e23fe {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.589477] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c959b13-a5f3-43f9-bf48-3ccb96c25eda {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.601186] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9774fca-19af-4b85-8618-baba09a698c3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.616247] env[69171]: DEBUG nova.compute.provider_tree [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 753.719138] env[69171]: DEBUG oslo_concurrency.lockutils [None req-067a5d5c-c7a6-4538-b7d4-7eeeea4e4f42 tempest-ServerAddressesTestJSON-111332568 tempest-ServerAddressesTestJSON-111332568-project-member] Lock "e0f59163-b386-45d3-87d7-1271deebe2f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.901s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 753.909310] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.149096] env[69171]: ERROR nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [req-2378da51-2cca-4eb7-8efc-2480ef420a15] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2378da51-2cca-4eb7-8efc-2480ef420a15"}]} [ 754.149673] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.190s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.153033] env[69171]: ERROR nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 754.153033] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Traceback (most recent call last): [ 754.153033] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 754.153033] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] yield [ 754.153033] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 754.153033] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] self.set_inventory_for_provider( [ 754.153033] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 754.153033] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2378da51-2cca-4eb7-8efc-2480ef420a15"}]} [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] During handling of the above exception, another exception occurred: [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Traceback (most recent call last): [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] with self.rt.instance_claim(context, instance, node, allocs, [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 754.153762] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] return f(*args, **kwargs) [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] self._update(elevated, cn) [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] self._update_to_placement(context, compute_node, startup) [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] return attempt.get(self._wrap_exception) [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] six.reraise(self.value[0], self.value[1], self.value[2]) [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] raise value [ 754.153991] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] self.reportclient.update_from_provider_tree( [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] with catch_all(pd.uuid): [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] self.gen.throw(typ, value, traceback) [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] raise exception.ResourceProviderSyncFailed() [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 754.154323] env[69171]: ERROR nova.compute.manager [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] [ 754.154569] env[69171]: DEBUG nova.compute.utils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 754.154569] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.357s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.156952] env[69171]: INFO nova.compute.claims [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.162560] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Build of instance 598fe41b-debf-4fbd-ac21-5cda132d2fea was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 754.162931] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 754.164286] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "refresh_cache-598fe41b-debf-4fbd-ac21-5cda132d2fea" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.164286] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquired lock "refresh_cache-598fe41b-debf-4fbd-ac21-5cda132d2fea" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.164286] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 754.221531] env[69171]: DEBUG nova.compute.manager [None req-7b994a0a-dc43-4089-b160-f31cc15302f9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 87075f00-32ea-4785-ace5-e88d07fe351c] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 754.412619] env[69171]: INFO nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 7a5215be-525e-4400-8972-4d6a949f5c24] Took 1.05 seconds to deallocate network for instance. [ 754.727941] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 754.733730] env[69171]: DEBUG nova.compute.manager [None req-7b994a0a-dc43-4089-b160-f31cc15302f9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 87075f00-32ea-4785-ace5-e88d07fe351c] Instance disappeared before build. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2440}} [ 754.935536] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Acquiring lock "221be95e-aac8-4551-8525-9745480ea8c6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.936243] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Lock "221be95e-aac8-4551-8525-9745480ea8c6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.972340] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.205797] env[69171]: DEBUG nova.scheduler.client.report [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 755.228068] env[69171]: DEBUG nova.scheduler.client.report [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 755.228068] env[69171]: DEBUG nova.compute.provider_tree [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 755.243952] env[69171]: DEBUG nova.scheduler.client.report [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 755.252444] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7b994a0a-dc43-4089-b160-f31cc15302f9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "87075f00-32ea-4785-ace5-e88d07fe351c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.975s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 755.273343] env[69171]: DEBUG nova.scheduler.client.report [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 755.451297] env[69171]: INFO nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Deleted allocations for instance 7a5215be-525e-4400-8972-4d6a949f5c24 [ 755.476608] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Releasing lock "refresh_cache-598fe41b-debf-4fbd-ac21-5cda132d2fea" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.476608] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 755.476608] env[69171]: DEBUG nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 755.476608] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 755.524168] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 755.758018] env[69171]: DEBUG nova.compute.manager [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 755.762898] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b718e016-4a81-4a5d-a008-8d38472ede3c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.774328] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e056ac92-ef00-4d75-9eee-0de6ca2aeb07 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.817671] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e02431d-6f24-4ec8-8ab6-d06507dc3cc9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.827275] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4599b189-bed2-46e9-b1d0-323dbc6f3a91 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.849124] env[69171]: DEBUG nova.compute.provider_tree [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 755.966301] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "7a5215be-525e-4400-8972-4d6a949f5c24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.077s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.028188] env[69171]: DEBUG nova.network.neutron [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.297374] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.385335] env[69171]: ERROR nova.scheduler.client.report [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [req-fd46cf74-815d-45a2-9e9a-f7d23c4f9d7d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fd46cf74-815d-45a2-9e9a-f7d23c4f9d7d"}]} [ 756.385928] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.232s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.386453] env[69171]: ERROR nova.compute.manager [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 756.386453] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Traceback (most recent call last): [ 756.386453] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 756.386453] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] yield [ 756.386453] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 756.386453] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] self.set_inventory_for_provider( [ 756.386453] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 756.386453] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fd46cf74-815d-45a2-9e9a-f7d23c4f9d7d"}]} [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] During handling of the above exception, another exception occurred: [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Traceback (most recent call last): [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] with self.rt.instance_claim(context, instance, node, allocs, [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 756.386857] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] return f(*args, **kwargs) [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] self._update(elevated, cn) [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] self._update_to_placement(context, compute_node, startup) [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] return attempt.get(self._wrap_exception) [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] raise value [ 756.388155] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] self.reportclient.update_from_provider_tree( [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] with catch_all(pd.uuid): [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] self.gen.throw(typ, value, traceback) [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] raise exception.ResourceProviderSyncFailed() [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 756.388943] env[69171]: ERROR nova.compute.manager [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] [ 756.389631] env[69171]: DEBUG nova.compute.utils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 756.389631] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.544s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.391223] env[69171]: INFO nova.compute.claims [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.396977] env[69171]: DEBUG nova.compute.manager [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Build of instance 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 756.397439] env[69171]: DEBUG nova.compute.manager [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 756.397669] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Acquiring lock "refresh_cache-377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.397817] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Acquired lock "refresh_cache-377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.399534] env[69171]: DEBUG nova.network.neutron [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 756.469567] env[69171]: DEBUG nova.compute.manager [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 756.532400] env[69171]: INFO nova.compute.manager [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 598fe41b-debf-4fbd-ac21-5cda132d2fea] Took 1.06 seconds to deallocate network for instance. [ 756.933895] env[69171]: DEBUG nova.network.neutron [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.006501] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.013244] env[69171]: DEBUG nova.network.neutron [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.436540] env[69171]: DEBUG nova.scheduler.client.report [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 757.453766] env[69171]: DEBUG nova.scheduler.client.report [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 757.453766] env[69171]: DEBUG nova.compute.provider_tree [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 757.465457] env[69171]: DEBUG nova.scheduler.client.report [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 757.490023] env[69171]: DEBUG nova.scheduler.client.report [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 757.518391] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Releasing lock "refresh_cache-377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.518652] env[69171]: DEBUG nova.compute.manager [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 757.518842] env[69171]: DEBUG nova.compute.manager [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 757.519028] env[69171]: DEBUG nova.network.neutron [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 757.549569] env[69171]: DEBUG nova.network.neutron [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 757.580523] env[69171]: INFO nova.scheduler.client.report [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Deleted allocations for instance 598fe41b-debf-4fbd-ac21-5cda132d2fea [ 757.849442] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e27806d-6183-403e-86b5-96363e918de7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.862614] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7a86fdd-f081-4706-9875-90495c158b8c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.896213] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed2a22c9-1d01-4762-acee-689a040b4f85 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.904814] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ffaf8d3-d48e-4bfd-8f9d-9dfe4b62f20b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.921395] env[69171]: DEBUG nova.compute.provider_tree [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 758.054300] env[69171]: DEBUG nova.network.neutron [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.092886] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0d7cf55a-f5e2-4525-8268-04c324fcbd52 tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "598fe41b-debf-4fbd-ac21-5cda132d2fea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.176s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.457262] env[69171]: ERROR nova.scheduler.client.report [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [req-0cac6f75-d785-4e90-966e-7ac026c3e09c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0cac6f75-d785-4e90-966e-7ac026c3e09c"}]} [ 758.458148] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.068s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.458841] env[69171]: ERROR nova.compute.manager [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 758.458841] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Traceback (most recent call last): [ 758.458841] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 758.458841] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] yield [ 758.458841] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 758.458841] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] self.set_inventory_for_provider( [ 758.458841] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 758.458841] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0cac6f75-d785-4e90-966e-7ac026c3e09c"}]} [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] During handling of the above exception, another exception occurred: [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Traceback (most recent call last): [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] with self.rt.instance_claim(context, instance, node, allocs, [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 758.459435] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] return f(*args, **kwargs) [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] self._update(elevated, cn) [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] self._update_to_placement(context, compute_node, startup) [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] return attempt.get(self._wrap_exception) [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] six.reraise(self.value[0], self.value[1], self.value[2]) [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] raise value [ 758.460185] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] self.reportclient.update_from_provider_tree( [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] with catch_all(pd.uuid): [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] self.gen.throw(typ, value, traceback) [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] raise exception.ResourceProviderSyncFailed() [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 758.462174] env[69171]: ERROR nova.compute.manager [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] [ 758.462511] env[69171]: DEBUG nova.compute.utils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 758.464021] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.426s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.465798] env[69171]: INFO nova.compute.claims [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 758.470557] env[69171]: DEBUG nova.compute.manager [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Build of instance d4220806-373f-45cc-82d0-cc70ca2c26ef was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 758.471097] env[69171]: DEBUG nova.compute.manager [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 758.471276] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Acquiring lock "refresh_cache-d4220806-373f-45cc-82d0-cc70ca2c26ef" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.471416] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Acquired lock "refresh_cache-d4220806-373f-45cc-82d0-cc70ca2c26ef" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.471578] env[69171]: DEBUG nova.network.neutron [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 758.521029] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Acquiring lock "47c2201c-a952-4474-88e1-2773f04494aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.521301] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Lock "47c2201c-a952-4474-88e1-2773f04494aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.556102] env[69171]: INFO nova.compute.manager [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] [instance: 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3] Took 1.04 seconds to deallocate network for instance. [ 758.599620] env[69171]: DEBUG nova.compute.manager [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 758.604059] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.605185] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.017089] env[69171]: DEBUG nova.network.neutron [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 759.142990] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.227800] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquiring lock "05eacaa7-1a14-4834-86ee-7f6dcaf21b85" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.228421] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "05eacaa7-1a14-4834-86ee-7f6dcaf21b85" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.289598] env[69171]: DEBUG nova.network.neutron [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.516722] env[69171]: DEBUG nova.scheduler.client.report [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 759.548261] env[69171]: DEBUG nova.scheduler.client.report [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 759.548261] env[69171]: DEBUG nova.compute.provider_tree [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 759.571647] env[69171]: DEBUG nova.scheduler.client.report [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 759.607987] env[69171]: DEBUG nova.scheduler.client.report [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 759.613673] env[69171]: INFO nova.scheduler.client.report [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Deleted allocations for instance 377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3 [ 759.792929] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Releasing lock "refresh_cache-d4220806-373f-45cc-82d0-cc70ca2c26ef" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.793188] env[69171]: DEBUG nova.compute.manager [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 759.793372] env[69171]: DEBUG nova.compute.manager [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 759.793555] env[69171]: DEBUG nova.network.neutron [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 759.852979] env[69171]: DEBUG nova.network.neutron [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 760.092815] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-960a73a6-6b78-4478-b0ea-8ae4bfed0058 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.102170] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19b59d4-43d9-4778-bb52-23913f90af81 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.141749] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e89158ff-e3ca-417d-b409-843f4b313611 tempest-TenantUsagesTestJSON-1189228773 tempest-TenantUsagesTestJSON-1189228773-project-member] Lock "377a8d4c-f905-4c09-ac2e-01c5f5bc5ad3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.627s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.147238] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38635f5-2dc2-40ea-9426-da2eea4e24c2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.157414] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b797437e-7bcd-422e-8d2f-b48f6173e84d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.177856] env[69171]: DEBUG nova.compute.provider_tree [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 760.358639] env[69171]: DEBUG nova.network.neutron [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.650201] env[69171]: DEBUG nova.compute.manager [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 760.712018] env[69171]: ERROR nova.scheduler.client.report [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [req-cb03bb1a-1277-4f70-9737-682fe26b9606] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-cb03bb1a-1277-4f70-9737-682fe26b9606"}]} [ 760.713141] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.713702] env[69171]: ERROR nova.compute.manager [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 760.713702] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Traceback (most recent call last): [ 760.713702] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 760.713702] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] yield [ 760.713702] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 760.713702] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] self.set_inventory_for_provider( [ 760.713702] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 760.713702] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-cb03bb1a-1277-4f70-9737-682fe26b9606"}]} [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] During handling of the above exception, another exception occurred: [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Traceback (most recent call last): [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] with self.rt.instance_claim(context, instance, node, allocs, [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 760.713951] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] return f(*args, **kwargs) [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] self._update(elevated, cn) [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] self._update_to_placement(context, compute_node, startup) [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] return attempt.get(self._wrap_exception) [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] six.reraise(self.value[0], self.value[1], self.value[2]) [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] raise value [ 760.714212] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] self.reportclient.update_from_provider_tree( [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] with catch_all(pd.uuid): [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] self.gen.throw(typ, value, traceback) [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] raise exception.ResourceProviderSyncFailed() [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 760.714525] env[69171]: ERROR nova.compute.manager [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] [ 760.717092] env[69171]: DEBUG nova.compute.utils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 760.717092] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.237s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 760.717286] env[69171]: DEBUG nova.objects.instance [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lazy-loading 'resources' on Instance uuid 8e6a9d73-6705-484c-82e0-70c6a6860211 {{(pid=69171) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 760.720226] env[69171]: DEBUG nova.compute.manager [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Build of instance 14990d60-4a6a-4896-816f-7bd7d5af19a5 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 760.720226] env[69171]: DEBUG nova.compute.manager [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 760.726844] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Acquiring lock "refresh_cache-14990d60-4a6a-4896-816f-7bd7d5af19a5" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 760.726844] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Acquired lock "refresh_cache-14990d60-4a6a-4896-816f-7bd7d5af19a5" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.726844] env[69171]: DEBUG nova.network.neutron [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 760.862575] env[69171]: INFO nova.compute.manager [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] [instance: d4220806-373f-45cc-82d0-cc70ca2c26ef] Took 1.07 seconds to deallocate network for instance. [ 761.187517] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.257391] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 761.285081] env[69171]: DEBUG nova.network.neutron [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 761.295278] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 761.295471] env[69171]: DEBUG nova.compute.provider_tree [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 761.306628] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "59600940-1b15-4192-906d-936d15b4377c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.307129] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "59600940-1b15-4192-906d-936d15b4377c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.315232] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 761.367134] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "55cf3ad8-d03c-4ac0-9c48-0a9382bc8628" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.367134] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "55cf3ad8-d03c-4ac0-9c48-0a9382bc8628" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.455697] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 761.535290] env[69171]: DEBUG nova.network.neutron [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.890635] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-254ab090-eaa4-4084-9bbe-42262dc56693 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.902217] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1e37c1-dbae-4c09-856e-103b02d2d904 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.944031] env[69171]: INFO nova.scheduler.client.report [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Deleted allocations for instance d4220806-373f-45cc-82d0-cc70ca2c26ef [ 761.950376] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2943af1c-2d01-4ca7-a6e5-05afaaf62cac {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.961362] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c66cb45-47ad-4ddf-a72c-fa5fb391a98b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.977739] env[69171]: DEBUG nova.compute.provider_tree [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 762.038442] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Releasing lock "refresh_cache-14990d60-4a6a-4896-816f-7bd7d5af19a5" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.038721] env[69171]: DEBUG nova.compute.manager [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 762.038947] env[69171]: DEBUG nova.compute.manager [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 762.039143] env[69171]: DEBUG nova.network.neutron [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 762.060945] env[69171]: DEBUG nova.network.neutron [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 762.456590] env[69171]: DEBUG oslo_concurrency.lockutils [None req-250053d1-ea19-4100-90df-6d7f1cba10b8 tempest-ServersTestManualDisk-1313495267 tempest-ServersTestManualDisk-1313495267-project-member] Lock "d4220806-373f-45cc-82d0-cc70ca2c26ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.020s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.502651] env[69171]: ERROR nova.scheduler.client.report [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [req-6d1db241-64b5-4d4f-bfcc-321c7eca36e0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6d1db241-64b5-4d4f-bfcc-321c7eca36e0"}]} [ 762.503043] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.503878] env[69171]: ERROR nova.compute.manager [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Setting instance vm_state to ERROR: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 762.503878] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Traceback (most recent call last): [ 762.503878] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 762.503878] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] yield [ 762.503878] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 762.503878] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self.set_inventory_for_provider( [ 762.503878] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 762.503878] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6d1db241-64b5-4d4f-bfcc-321c7eca36e0"}]} [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] During handling of the above exception, another exception occurred: [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Traceback (most recent call last): [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self._delete_instance(context, instance, bdms) [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 762.504286] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self._complete_deletion(context, instance) [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self._update_resource_tracker(context, instance) [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self.rt.update_usage(context, instance, instance.node) [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] return f(*args, **kwargs) [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self._update(context.elevated(), self.compute_nodes[nodename]) [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self._update_to_placement(context, compute_node, startup) [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 762.504523] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] return attempt.get(self._wrap_exception) [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] six.reraise(self.value[0], self.value[1], self.value[2]) [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] raise value [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self.reportclient.update_from_provider_tree( [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] with catch_all(pd.uuid): [ 762.505524] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 762.505863] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] self.gen.throw(typ, value, traceback) [ 762.505863] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 762.505863] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] raise exception.ResourceProviderSyncFailed() [ 762.505863] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 762.505863] env[69171]: ERROR nova.compute.manager [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] [ 762.506203] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.171s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.507619] env[69171]: INFO nova.compute.claims [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 762.563342] env[69171]: DEBUG nova.network.neutron [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.947862] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Acquiring lock "62fecdeb-6657-4fc5-934c-5c459e7287d7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.949406] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Lock "62fecdeb-6657-4fc5-934c-5c459e7287d7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.962264] env[69171]: DEBUG nova.compute.manager [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 763.012778] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "8e6a9d73-6705-484c-82e0-70c6a6860211" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 22.210s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.067301] env[69171]: INFO nova.compute.manager [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] [instance: 14990d60-4a6a-4896-816f-7bd7d5af19a5] Took 1.03 seconds to deallocate network for instance. [ 763.497577] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.556434] env[69171]: DEBUG nova.scheduler.client.report [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 763.584792] env[69171]: DEBUG nova.scheduler.client.report [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 763.585347] env[69171]: DEBUG nova.compute.provider_tree [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 763.618302] env[69171]: DEBUG nova.scheduler.client.report [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 763.654456] env[69171]: DEBUG nova.scheduler.client.report [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 764.089304] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-835f2565-0263-4689-b693-51e6cfbec426 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.105771] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cee5a30b-1e50-46bc-803a-6e7329139be6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.163714] env[69171]: INFO nova.scheduler.client.report [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Deleted allocations for instance 14990d60-4a6a-4896-816f-7bd7d5af19a5 [ 764.170666] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f63b316-389d-47c9-8cb6-149420f4cf70 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.188361] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5499c50-a8b5-4bd3-815c-d1591070fb79 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.212198] env[69171]: DEBUG nova.compute.provider_tree [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 764.483875] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Acquiring lock "6664e9ec-4c43-464d-9117-92967031fe55" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.485271] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Lock "6664e9ec-4c43-464d-9117-92967031fe55" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.535218] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 764.678446] env[69171]: DEBUG oslo_concurrency.lockutils [None req-6db678ec-af76-4863-817e-fe980cc24424 tempest-AttachInterfacesUnderV243Test-736327661 tempest-AttachInterfacesUnderV243Test-736327661-project-member] Lock "14990d60-4a6a-4896-816f-7bd7d5af19a5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.111s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.745320] env[69171]: ERROR nova.scheduler.client.report [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [req-1bef70f4-feaf-4fe5-8c3a-ef2ee3dfc884] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1bef70f4-feaf-4fe5-8c3a-ef2ee3dfc884"}]} [ 764.745722] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.240s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.746309] env[69171]: ERROR nova.compute.manager [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 764.746309] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Traceback (most recent call last): [ 764.746309] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 764.746309] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] yield [ 764.746309] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 764.746309] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] self.set_inventory_for_provider( [ 764.746309] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 764.746309] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1bef70f4-feaf-4fe5-8c3a-ef2ee3dfc884"}]} [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] During handling of the above exception, another exception occurred: [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Traceback (most recent call last): [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] with self.rt.instance_claim(context, instance, node, allocs, [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 764.746553] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] return f(*args, **kwargs) [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] self._update(elevated, cn) [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] self._update_to_placement(context, compute_node, startup) [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] return attempt.get(self._wrap_exception) [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] six.reraise(self.value[0], self.value[1], self.value[2]) [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] raise value [ 764.746822] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] self.reportclient.update_from_provider_tree( [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] with catch_all(pd.uuid): [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] self.gen.throw(typ, value, traceback) [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] raise exception.ResourceProviderSyncFailed() [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 764.747221] env[69171]: ERROR nova.compute.manager [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] [ 764.747731] env[69171]: DEBUG nova.compute.utils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 764.754560] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.660s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.754837] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.756941] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.486s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.762158] env[69171]: INFO nova.compute.claims [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 764.764702] env[69171]: DEBUG nova.compute.manager [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Build of instance 17c421bd-a02f-41a6-9074-8c143e423c79 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 764.765366] env[69171]: DEBUG nova.compute.manager [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 764.765712] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Acquiring lock "refresh_cache-17c421bd-a02f-41a6-9074-8c143e423c79" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.766099] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Acquired lock "refresh_cache-17c421bd-a02f-41a6-9074-8c143e423c79" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.768566] env[69171]: DEBUG nova.network.neutron [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 764.797846] env[69171]: INFO nova.scheduler.client.report [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Deleted allocations for instance fa5997fa-817c-42b6-a7fa-b3b85a109b6d [ 765.181619] env[69171]: DEBUG nova.compute.manager [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 765.306796] env[69171]: DEBUG nova.network.neutron [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 765.312363] env[69171]: DEBUG oslo_concurrency.lockutils [None req-75f70d50-1bc4-42e7-9c0e-72edb5b187cf tempest-DeleteServersAdminTestJSON-1424598227 tempest-DeleteServersAdminTestJSON-1424598227-project-member] Lock "fa5997fa-817c-42b6-a7fa-b3b85a109b6d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.051s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.437857] env[69171]: DEBUG nova.network.neutron [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.712295] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.799378] env[69171]: DEBUG nova.scheduler.client.report [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 765.814677] env[69171]: DEBUG nova.scheduler.client.report [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 765.814677] env[69171]: DEBUG nova.compute.provider_tree [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 765.839497] env[69171]: DEBUG nova.scheduler.client.report [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 765.866407] env[69171]: DEBUG nova.scheduler.client.report [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 765.938899] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Releasing lock "refresh_cache-17c421bd-a02f-41a6-9074-8c143e423c79" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.939110] env[69171]: DEBUG nova.compute.manager [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 765.939335] env[69171]: DEBUG nova.compute.manager [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 765.940028] env[69171]: DEBUG nova.network.neutron [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 765.980262] env[69171]: DEBUG nova.network.neutron [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 766.308331] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d2d79d-d53b-4636-a635-c016fe33e761 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.323549] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf90461-092a-467d-9adc-951dd2e7a862 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.360224] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa42af5f-697c-4aa6-b3fe-673d2449affc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.369981] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c5cde7a-0667-45d0-92d1-d54b2acab691 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.393760] env[69171]: DEBUG nova.compute.provider_tree [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 766.486207] env[69171]: DEBUG nova.network.neutron [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.929325] env[69171]: ERROR nova.scheduler.client.report [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [req-ab4243bd-c089-4221-b95a-6d828bf76acc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ab4243bd-c089-4221-b95a-6d828bf76acc"}]} [ 766.929703] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.930291] env[69171]: ERROR nova.compute.manager [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 766.930291] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Traceback (most recent call last): [ 766.930291] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 766.930291] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] yield [ 766.930291] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 766.930291] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] self.set_inventory_for_provider( [ 766.930291] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 766.930291] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ab4243bd-c089-4221-b95a-6d828bf76acc"}]} [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] During handling of the above exception, another exception occurred: [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Traceback (most recent call last): [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] with self.rt.instance_claim(context, instance, node, allocs, [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 766.930519] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] return f(*args, **kwargs) [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] self._update(elevated, cn) [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] self._update_to_placement(context, compute_node, startup) [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] return attempt.get(self._wrap_exception) [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] six.reraise(self.value[0], self.value[1], self.value[2]) [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] raise value [ 766.930775] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] self.reportclient.update_from_provider_tree( [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] with catch_all(pd.uuid): [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] self.gen.throw(typ, value, traceback) [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] raise exception.ResourceProviderSyncFailed() [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 766.931166] env[69171]: ERROR nova.compute.manager [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] [ 766.935089] env[69171]: DEBUG nova.compute.utils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 766.936570] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.420s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.938690] env[69171]: INFO nova.compute.claims [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.942961] env[69171]: DEBUG nova.compute.manager [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Build of instance 155b53f6-a42a-48ad-bbfb-a99fc9eba789 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 766.943446] env[69171]: DEBUG nova.compute.manager [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 766.943680] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "refresh_cache-155b53f6-a42a-48ad-bbfb-a99fc9eba789" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.944672] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquired lock "refresh_cache-155b53f6-a42a-48ad-bbfb-a99fc9eba789" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.944889] env[69171]: DEBUG nova.network.neutron [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 766.991741] env[69171]: INFO nova.compute.manager [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] [instance: 17c421bd-a02f-41a6-9074-8c143e423c79] Took 1.05 seconds to deallocate network for instance. [ 767.470184] env[69171]: DEBUG nova.network.neutron [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 767.537838] env[69171]: DEBUG nova.network.neutron [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.860495] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Acquiring lock "0c0f5347-23e3-4e69-8b9d-56eeebf39a6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.860495] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Lock "0c0f5347-23e3-4e69-8b9d-56eeebf39a6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 767.974023] env[69171]: DEBUG nova.scheduler.client.report [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 767.994669] env[69171]: DEBUG nova.scheduler.client.report [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 767.994918] env[69171]: DEBUG nova.compute.provider_tree [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 768.009352] env[69171]: DEBUG nova.scheduler.client.report [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 768.025780] env[69171]: INFO nova.scheduler.client.report [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Deleted allocations for instance 17c421bd-a02f-41a6-9074-8c143e423c79 [ 768.032819] env[69171]: DEBUG nova.scheduler.client.report [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 768.045415] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Releasing lock "refresh_cache-155b53f6-a42a-48ad-bbfb-a99fc9eba789" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.045415] env[69171]: DEBUG nova.compute.manager [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 768.045415] env[69171]: DEBUG nova.compute.manager [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 768.045415] env[69171]: DEBUG nova.network.neutron [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 768.089165] env[69171]: DEBUG nova.network.neutron [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 768.442195] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Acquiring lock "aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.442195] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Lock "aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.515855] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79298c82-2d00-49ee-b5a7-dae61cd8735d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.525377] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-176f601d-1989-4e7b-88fb-b829a79dcc32 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.560721] env[69171]: DEBUG oslo_concurrency.lockutils [None req-73d32502-f31f-4831-8534-abb214923b89 tempest-InstanceActionsTestJSON-494896742 tempest-InstanceActionsTestJSON-494896742-project-member] Lock "17c421bd-a02f-41a6-9074-8c143e423c79" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.814s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.562884] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95795671-8c7e-46c9-ac7e-bf73ffc96da3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.572227] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8e6c7a-62bf-47d7-bb79-6d4e3ff0cb6a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.592349] env[69171]: DEBUG nova.compute.provider_tree [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 768.593990] env[69171]: DEBUG nova.network.neutron [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.069794] env[69171]: DEBUG nova.compute.manager [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 769.103454] env[69171]: INFO nova.compute.manager [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: 155b53f6-a42a-48ad-bbfb-a99fc9eba789] Took 1.06 seconds to deallocate network for instance. [ 769.134598] env[69171]: ERROR nova.scheduler.client.report [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [req-3ab6d314-52e3-4a05-893b-a5afa4bb5a3a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3ab6d314-52e3-4a05-893b-a5afa4bb5a3a"}]} [ 769.135049] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.198s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.135896] env[69171]: ERROR nova.compute.manager [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 769.135896] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Traceback (most recent call last): [ 769.135896] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 769.135896] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] yield [ 769.135896] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 769.135896] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] self.set_inventory_for_provider( [ 769.135896] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 769.135896] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3ab6d314-52e3-4a05-893b-a5afa4bb5a3a"}]} [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] During handling of the above exception, another exception occurred: [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Traceback (most recent call last): [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] with self.rt.instance_claim(context, instance, node, allocs, [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 769.136372] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] return f(*args, **kwargs) [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] self._update(elevated, cn) [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] self._update_to_placement(context, compute_node, startup) [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] return attempt.get(self._wrap_exception) [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] six.reraise(self.value[0], self.value[1], self.value[2]) [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] raise value [ 769.138186] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] self.reportclient.update_from_provider_tree( [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] with catch_all(pd.uuid): [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] self.gen.throw(typ, value, traceback) [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] raise exception.ResourceProviderSyncFailed() [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 769.141128] env[69171]: ERROR nova.compute.manager [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] [ 769.142207] env[69171]: DEBUG nova.compute.utils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 769.144767] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 18.017s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.145101] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 769.145324] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=69171) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 769.146218] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.323s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 769.148343] env[69171]: INFO nova.compute.claims [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 769.152548] env[69171]: DEBUG nova.compute.manager [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Build of instance ea94a5c3-226a-4866-8d96-2e5c341d4636 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 769.155699] env[69171]: DEBUG nova.compute.manager [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 769.155699] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquiring lock "refresh_cache-ea94a5c3-226a-4866-8d96-2e5c341d4636" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 769.155699] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Acquired lock "refresh_cache-ea94a5c3-226a-4866-8d96-2e5c341d4636" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.156565] env[69171]: DEBUG nova.network.neutron [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 769.158203] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbd3891-d809-4a89-a10a-4d7fab5bf56d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.173713] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c243a8b-9fdd-4334-bb41-766985c9cc28 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.195875] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9edde2d3-cb43-49ba-b13c-6461d659edb3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.206685] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9456d1c-3504-4bca-b187-cc800a3f9dde {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.244777] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180016MB free_disk=0GB free_vcpus=48 pci_devices=None {{(pid=69171) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 769.244924] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.602368] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.686876] env[69171]: DEBUG nova.network.neutron [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 769.816356] env[69171]: DEBUG nova.network.neutron [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.139364] env[69171]: INFO nova.scheduler.client.report [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Deleted allocations for instance 155b53f6-a42a-48ad-bbfb-a99fc9eba789 [ 770.219965] env[69171]: DEBUG nova.scheduler.client.report [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 770.256486] env[69171]: DEBUG nova.scheduler.client.report [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 770.256766] env[69171]: DEBUG nova.compute.provider_tree [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 770.281485] env[69171]: DEBUG nova.scheduler.client.report [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 770.309674] env[69171]: DEBUG nova.scheduler.client.report [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 770.317598] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Releasing lock "refresh_cache-ea94a5c3-226a-4866-8d96-2e5c341d4636" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.317598] env[69171]: DEBUG nova.compute.manager [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 770.317598] env[69171]: DEBUG nova.compute.manager [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 770.317598] env[69171]: DEBUG nova.network.neutron [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 770.346722] env[69171]: DEBUG nova.network.neutron [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 770.635492] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b2dc39b-cd23-4465-b275-a79eef354a96 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.644536] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08495968-ac2f-417d-8f3d-218ccecb18a4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.649680] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8b83b162-c605-4f6b-90ab-b62593d2f5e1 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "155b53f6-a42a-48ad-bbfb-a99fc9eba789" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.958s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.680146] env[69171]: DEBUG nova.compute.manager [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 770.683128] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bbbd56e-fbe6-4644-8b3b-f05b20e81f25 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.694706] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f7bfd34-ff83-4e99-ac0c-8fbcaecd6029 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.708279] env[69171]: DEBUG nova.compute.provider_tree [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 770.849411] env[69171]: DEBUG nova.network.neutron [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.209618] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.237238] env[69171]: ERROR nova.scheduler.client.report [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [req-0670023b-8c58-4421-b3c9-2f4e326bb1bf] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0670023b-8c58-4421-b3c9-2f4e326bb1bf"}]} [ 771.237619] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.091s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.238632] env[69171]: ERROR nova.compute.manager [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 771.238632] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Traceback (most recent call last): [ 771.238632] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 771.238632] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] yield [ 771.238632] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 771.238632] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] self.set_inventory_for_provider( [ 771.238632] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 771.238632] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0670023b-8c58-4421-b3c9-2f4e326bb1bf"}]} [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] During handling of the above exception, another exception occurred: [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Traceback (most recent call last): [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] with self.rt.instance_claim(context, instance, node, allocs, [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 771.238894] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] return f(*args, **kwargs) [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] self._update(elevated, cn) [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] self._update_to_placement(context, compute_node, startup) [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] return attempt.get(self._wrap_exception) [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] six.reraise(self.value[0], self.value[1], self.value[2]) [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] raise value [ 771.239183] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] self.reportclient.update_from_provider_tree( [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] with catch_all(pd.uuid): [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] self.gen.throw(typ, value, traceback) [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] raise exception.ResourceProviderSyncFailed() [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 771.239574] env[69171]: ERROR nova.compute.manager [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] [ 771.239979] env[69171]: DEBUG nova.compute.utils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 771.241888] env[69171]: DEBUG nova.compute.manager [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Build of instance d0facc29-338c-4d0d-a9f3-a37781c19b68 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 771.242365] env[69171]: DEBUG nova.compute.manager [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 771.242626] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquiring lock "refresh_cache-d0facc29-338c-4d0d-a9f3-a37781c19b68" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 771.242804] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquired lock "refresh_cache-d0facc29-338c-4d0d-a9f3-a37781c19b68" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 771.242997] env[69171]: DEBUG nova.network.neutron [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 771.245072] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.947s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.246840] env[69171]: INFO nova.compute.claims [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 771.354562] env[69171]: INFO nova.compute.manager [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] [instance: ea94a5c3-226a-4866-8d96-2e5c341d4636] Took 1.04 seconds to deallocate network for instance. [ 771.463325] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "f8b20421-15d3-46f3-a504-172d044ff4d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.463807] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.463807] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "f8b20421-15d3-46f3-a504-172d044ff4d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.464766] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.464766] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.467779] env[69171]: INFO nova.compute.manager [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Terminating instance [ 771.471277] env[69171]: DEBUG nova.compute.manager [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Start destroying the instance on the hypervisor. {{(pid=69171) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3151}} [ 771.472025] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Destroying instance {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1142}} [ 771.473802] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-072a995b-cd45-42e9-bd59-40db31d437c8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.485129] env[69171]: DEBUG nova.virt.vmwareapi.vm_util [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Powering off the VM {{(pid=69171) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1502}} [ 771.485129] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-74727452-053b-4f06-acc6-89f6ff53aa4e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.493394] env[69171]: DEBUG oslo_vmware.api [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for the task: (returnval){ [ 771.493394] env[69171]: value = "task-4235950" [ 771.493394] env[69171]: _type = "Task" [ 771.493394] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.504562] env[69171]: DEBUG oslo_vmware.api [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Task: {'id': task-4235950, 'name': PowerOffVM_Task} progress is 0%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.774648] env[69171]: DEBUG nova.network.neutron [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 771.990692] env[69171]: DEBUG nova.network.neutron [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.004458] env[69171]: DEBUG oslo_vmware.exceptions [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Fault QuestionPending not matched. {{(pid=69171) get_fault_class /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/exceptions.py:290}} [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Destroy instance failed: oslo_vmware.exceptions.VimFaultException: The operation cannot be allowed at the current time because the virtual machine has a question pending: [ 772.005919] env[69171]: 'msg.hbacommon.outofspace:There is no more space for virtual disk 'f8b20421-15d3-46f3-a504-172d044ff4d3.vmdk'. You might be able to continue this session by freeing disk space on the relevant volume, and clicking Retry. Click Cancel to terminate this session. [ 772.005919] env[69171]: '. [ 772.005919] env[69171]: Faults: ['QuestionPending'] [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Traceback (most recent call last): [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 1090, in _destroy_instance [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] vm_util.power_off_instance(self._session, instance, vm_ref) [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1506, in power_off_instance [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] session._wait_for_task(poweroff_task) [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 157, in _wait_for_task [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return self.wait_for_task(task_ref) [ 772.005919] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 398, in wait_for_task [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return evt.wait() [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] result = hub.switch() [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] return self.greenlet.switch() [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 75, in _inner [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] self.f(*self.args, **self.kw) [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 448, in _poll_task [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] raise exceptions.translate_fault(task_info.error) [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] oslo_vmware.exceptions.VimFaultException: The operation cannot be allowed at the current time because the virtual machine has a question pending: [ 772.006398] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] 'msg.hbacommon.outofspace:There is no more space for virtual disk 'f8b20421-15d3-46f3-a504-172d044ff4d3.vmdk'. You might be able to continue this session by freeing disk space on the relevant volume, and clicking Retry. Click Cancel to terminate this session. [ 772.006708] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] '. [ 772.006708] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Faults: ['QuestionPending'] [ 772.006708] env[69171]: ERROR nova.virt.vmwareapi.vmops [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] [ 772.006708] env[69171]: DEBUG nova.virt.vmwareapi.vmops [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Instance destroyed {{(pid=69171) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1144}} [ 772.006708] env[69171]: INFO nova.compute.manager [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Took 0.53 seconds to destroy the instance on the hypervisor. [ 772.006708] env[69171]: DEBUG oslo.service.loopingcall [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=69171) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 772.008205] env[69171]: DEBUG nova.compute.manager [-] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 772.008250] env[69171]: DEBUG nova.network.neutron [-] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 772.284932] env[69171]: DEBUG nova.scheduler.client.report [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 772.310692] env[69171]: DEBUG nova.scheduler.client.report [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 772.310983] env[69171]: DEBUG nova.compute.provider_tree [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 772.330204] env[69171]: DEBUG nova.scheduler.client.report [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 772.355667] env[69171]: DEBUG nova.scheduler.client.report [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 772.394407] env[69171]: INFO nova.scheduler.client.report [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Deleted allocations for instance ea94a5c3-226a-4866-8d96-2e5c341d4636 [ 772.500645] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Releasing lock "refresh_cache-d0facc29-338c-4d0d-a9f3-a37781c19b68" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.500921] env[69171]: DEBUG nova.compute.manager [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 772.501400] env[69171]: DEBUG nova.compute.manager [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 772.501400] env[69171]: DEBUG nova.network.neutron [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 772.527878] env[69171]: DEBUG nova.network.neutron [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 772.613109] env[69171]: DEBUG nova.compute.manager [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Received event network-vif-deleted-8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7 {{(pid=69171) external_instance_event /opt/stack/nova/nova/compute/manager.py:11132}} [ 772.613325] env[69171]: INFO nova.compute.manager [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Neutron deleted interface 8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7; detaching it from the instance and deleting it from the info cache [ 772.613489] env[69171]: DEBUG nova.network.neutron [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.778646] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e303223c-5988-4cdf-b778-79c2534ea25e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.790868] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c93af055-c144-4261-8225-8f308fc9d090 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.834247] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31e93aa-c0b6-48ef-b87a-b5f3734a97a7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.840910] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb99aea-2f42-4d0b-becd-0066f7e319ea {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.856787] env[69171]: DEBUG nova.compute.provider_tree [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 772.908073] env[69171]: DEBUG oslo_concurrency.lockutils [None req-982aac5f-f885-4a00-a095-1f366e2e723c tempest-VolumesAdminNegativeTest-1110994680 tempest-VolumesAdminNegativeTest-1110994680-project-member] Lock "ea94a5c3-226a-4866-8d96-2e5c341d4636" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.503s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.006893] env[69171]: DEBUG nova.network.neutron [-] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.033405] env[69171]: DEBUG nova.network.neutron [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.121896] env[69171]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c0055abe-742b-4bd2-a4ec-c1d8938e02b6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.130188] env[69171]: DEBUG oslo_concurrency.lockutils [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] Acquiring lock "f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.370214] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Acquiring lock "54d5f178-7fc5-48ce-be89-c783f419e581" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 773.370576] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Lock "54d5f178-7fc5-48ce-be89-c783f419e581" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.387671] env[69171]: ERROR nova.scheduler.client.report [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [req-12197d44-f305-4609-ae8f-ed5507b24c6a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-12197d44-f305-4609-ae8f-ed5507b24c6a"}]} [ 773.387671] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.143s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 773.389733] env[69171]: ERROR nova.compute.manager [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 773.389733] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Traceback (most recent call last): [ 773.389733] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 773.389733] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] yield [ 773.389733] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 773.389733] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] self.set_inventory_for_provider( [ 773.389733] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 773.389733] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-12197d44-f305-4609-ae8f-ed5507b24c6a"}]} [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] During handling of the above exception, another exception occurred: [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Traceback (most recent call last): [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] with self.rt.instance_claim(context, instance, node, allocs, [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 773.390086] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] return f(*args, **kwargs) [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] self._update(elevated, cn) [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] self._update_to_placement(context, compute_node, startup) [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] return attempt.get(self._wrap_exception) [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] raise value [ 773.390449] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] self.reportclient.update_from_provider_tree( [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] with catch_all(pd.uuid): [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] self.gen.throw(typ, value, traceback) [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] raise exception.ResourceProviderSyncFailed() [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 773.390968] env[69171]: ERROR nova.compute.manager [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] [ 773.391589] env[69171]: DEBUG nova.compute.utils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 773.392591] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.386s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 773.394170] env[69171]: INFO nova.compute.claims [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 773.396844] env[69171]: DEBUG nova.compute.manager [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Build of instance 4c5586e7-333c-44b0-aef9-4d1b7988826d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 773.401431] env[69171]: DEBUG nova.compute.manager [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 773.405018] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Acquiring lock "refresh_cache-4c5586e7-333c-44b0-aef9-4d1b7988826d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.405018] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Acquired lock "refresh_cache-4c5586e7-333c-44b0-aef9-4d1b7988826d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.405018] env[69171]: DEBUG nova.network.neutron [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 773.410871] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 773.509736] env[69171]: INFO nova.compute.manager [-] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Took 1.50 seconds to deallocate network for instance. [ 773.538454] env[69171]: INFO nova.compute.manager [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: d0facc29-338c-4d0d-a9f3-a37781c19b68] Took 1.04 seconds to deallocate network for instance. [ 773.947520] env[69171]: DEBUG nova.network.neutron [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 773.955307] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.019595] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 774.182147] env[69171]: DEBUG nova.network.neutron [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.435525] env[69171]: DEBUG nova.scheduler.client.report [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 774.459112] env[69171]: DEBUG nova.scheduler.client.report [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 774.459112] env[69171]: DEBUG nova.compute.provider_tree [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 774.476244] env[69171]: DEBUG nova.scheduler.client.report [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 774.498641] env[69171]: DEBUG nova.scheduler.client.report [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 774.589685] env[69171]: INFO nova.scheduler.client.report [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Deleted allocations for instance d0facc29-338c-4d0d-a9f3-a37781c19b68 [ 774.687162] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Releasing lock "refresh_cache-4c5586e7-333c-44b0-aef9-4d1b7988826d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.687162] env[69171]: DEBUG nova.compute.manager [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 774.687162] env[69171]: DEBUG nova.compute.manager [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 774.687162] env[69171]: DEBUG nova.network.neutron [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 774.721629] env[69171]: DEBUG nova.network.neutron [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 774.885349] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c007fb-7cee-43c6-a405-248dcb431b62 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.893980] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcfc873-ac2e-4e85-b7d9-9e6c1705bca7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.726101] env[69171]: DEBUG oslo_concurrency.lockutils [None req-86c736ae-fff4-4bc1-a51c-a3c7ac433013 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "d0facc29-338c-4d0d-a9f3-a37781c19b68" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 36.598s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.726411] env[69171]: DEBUG nova.network.neutron [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.732156] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b9b22e-e6de-4415-9948-de7baa071e3c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.738805] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e17185ed-1c2d-4fd4-970e-201fabcff6a1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.756676] env[69171]: DEBUG nova.compute.provider_tree [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 776.229654] env[69171]: INFO nova.compute.manager [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] [instance: 4c5586e7-333c-44b0-aef9-4d1b7988826d] Took 1.54 seconds to deallocate network for instance. [ 776.236341] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 776.287554] env[69171]: ERROR nova.scheduler.client.report [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [req-a7f359ac-f613-4436-baf7-6d9415b1e597] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a7f359ac-f613-4436-baf7-6d9415b1e597"}]} [ 776.287910] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.895s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.288564] env[69171]: ERROR nova.compute.manager [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 776.288564] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Traceback (most recent call last): [ 776.288564] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 776.288564] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] yield [ 776.288564] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 776.288564] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] self.set_inventory_for_provider( [ 776.288564] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 776.288564] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a7f359ac-f613-4436-baf7-6d9415b1e597"}]} [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] During handling of the above exception, another exception occurred: [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Traceback (most recent call last): [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] with self.rt.instance_claim(context, instance, node, allocs, [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 776.288813] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] return f(*args, **kwargs) [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] self._update(elevated, cn) [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] self._update_to_placement(context, compute_node, startup) [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] return attempt.get(self._wrap_exception) [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] six.reraise(self.value[0], self.value[1], self.value[2]) [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] raise value [ 776.289085] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] self.reportclient.update_from_provider_tree( [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] with catch_all(pd.uuid): [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] self.gen.throw(typ, value, traceback) [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] raise exception.ResourceProviderSyncFailed() [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 776.289399] env[69171]: ERROR nova.compute.manager [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] [ 776.289717] env[69171]: DEBUG nova.compute.utils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 776.292255] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.148s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.294889] env[69171]: INFO nova.compute.claims [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.298139] env[69171]: DEBUG nova.compute.manager [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Build of instance 6cdbe208-0eb2-4493-a477-daeb70739f95 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 776.300017] env[69171]: DEBUG nova.compute.manager [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 776.300017] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "refresh_cache-6cdbe208-0eb2-4493-a477-daeb70739f95" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 776.300017] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquired lock "refresh_cache-6cdbe208-0eb2-4493-a477-daeb70739f95" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 776.300017] env[69171]: DEBUG nova.network.neutron [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 776.777883] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 776.854886] env[69171]: DEBUG nova.network.neutron [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 776.979220] env[69171]: DEBUG nova.network.neutron [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 777.291289] env[69171]: INFO nova.scheduler.client.report [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Deleted allocations for instance 4c5586e7-333c-44b0-aef9-4d1b7988826d [ 777.337970] env[69171]: DEBUG nova.scheduler.client.report [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 777.358524] env[69171]: DEBUG nova.scheduler.client.report [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 777.359415] env[69171]: DEBUG nova.compute.provider_tree [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 777.381872] env[69171]: DEBUG nova.scheduler.client.report [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 777.412152] env[69171]: DEBUG nova.scheduler.client.report [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 777.488519] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Releasing lock "refresh_cache-6cdbe208-0eb2-4493-a477-daeb70739f95" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 777.488742] env[69171]: DEBUG nova.compute.manager [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 777.488952] env[69171]: DEBUG nova.compute.manager [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 777.491084] env[69171]: DEBUG nova.network.neutron [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 777.544221] env[69171]: DEBUG nova.network.neutron [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 777.809826] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8836fe45-5b67-4c4d-b99a-30fbbfa2e9b3 tempest-InstanceActionsNegativeTestJSON-1990988138 tempest-InstanceActionsNegativeTestJSON-1990988138-project-member] Lock "4c5586e7-333c-44b0-aef9-4d1b7988826d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 33.570s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.820989] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bc9e1e7-f456-4a8a-a9bb-991ac5f1c2ab {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.835428] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3934ff7-1658-46a3-a9f0-e6d81532555e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.877950] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a27489e9-fdae-468a-8faa-34db7290f3f6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.886996] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159e2b9b-1a8b-40ee-94b7-81a63e022099 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 777.904605] env[69171]: DEBUG nova.compute.provider_tree [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 778.049153] env[69171]: DEBUG nova.network.neutron [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.314276] env[69171]: DEBUG nova.compute.manager [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 778.432683] env[69171]: ERROR nova.scheduler.client.report [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [req-0982e687-4ce0-4dee-8408-4662c7f1f422] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0982e687-4ce0-4dee-8408-4662c7f1f422"}]} [ 778.433082] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.142s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.433743] env[69171]: ERROR nova.compute.manager [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 778.433743] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Traceback (most recent call last): [ 778.433743] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 778.433743] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] yield [ 778.433743] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 778.433743] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] self.set_inventory_for_provider( [ 778.433743] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 778.433743] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0982e687-4ce0-4dee-8408-4662c7f1f422"}]} [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] During handling of the above exception, another exception occurred: [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Traceback (most recent call last): [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] with self.rt.instance_claim(context, instance, node, allocs, [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 778.433972] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] return f(*args, **kwargs) [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] self._update(elevated, cn) [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] self._update_to_placement(context, compute_node, startup) [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] return attempt.get(self._wrap_exception) [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] six.reraise(self.value[0], self.value[1], self.value[2]) [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] raise value [ 778.434263] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] self.reportclient.update_from_provider_tree( [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] with catch_all(pd.uuid): [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] self.gen.throw(typ, value, traceback) [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] raise exception.ResourceProviderSyncFailed() [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 778.434822] env[69171]: ERROR nova.compute.manager [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] [ 778.435168] env[69171]: DEBUG nova.compute.utils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 778.435826] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.248s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 778.437277] env[69171]: INFO nova.compute.claims [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 778.440972] env[69171]: DEBUG nova.compute.manager [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Build of instance 9bd1c816-5248-4e48-89f8-f3b999d50bff was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 778.441467] env[69171]: DEBUG nova.compute.manager [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 778.441696] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquiring lock "refresh_cache-9bd1c816-5248-4e48-89f8-f3b999d50bff" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 778.441842] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Acquired lock "refresh_cache-9bd1c816-5248-4e48-89f8-f3b999d50bff" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 778.442026] env[69171]: DEBUG nova.network.neutron [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 778.557131] env[69171]: INFO nova.compute.manager [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6cdbe208-0eb2-4493-a477-daeb70739f95] Took 1.07 seconds to deallocate network for instance. [ 778.846139] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 778.971244] env[69171]: DEBUG nova.network.neutron [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.107035] env[69171]: DEBUG nova.network.neutron [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.479419] env[69171]: DEBUG nova.scheduler.client.report [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 779.505318] env[69171]: DEBUG nova.scheduler.client.report [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 779.505538] env[69171]: DEBUG nova.compute.provider_tree [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 779.526326] env[69171]: DEBUG nova.scheduler.client.report [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 779.551943] env[69171]: DEBUG nova.scheduler.client.report [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 779.568873] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Acquiring lock "478e1a26-101c-48a4-bcc2-a103f59962f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.569149] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Lock "478e1a26-101c-48a4-bcc2-a103f59962f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.611193] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Releasing lock "refresh_cache-9bd1c816-5248-4e48-89f8-f3b999d50bff" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 779.611440] env[69171]: DEBUG nova.compute.manager [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 779.611620] env[69171]: DEBUG nova.compute.manager [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 779.611780] env[69171]: DEBUG nova.network.neutron [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 779.615323] env[69171]: INFO nova.scheduler.client.report [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Deleted allocations for instance 6cdbe208-0eb2-4493-a477-daeb70739f95 [ 779.668168] env[69171]: DEBUG nova.network.neutron [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 779.889939] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Acquiring lock "ed037ce0-072f-425a-97fa-5fb0f941268b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.889939] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Lock "ed037ce0-072f-425a-97fa-5fb0f941268b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.986210] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79a2d5e4-e537-4792-a693-ac8f94a14248 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 779.996169] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6c8846-5aa4-4d60-8f30-85f5b4c1047e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.034776] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add37b84-7d10-48a5-9de3-f7bfb7697057 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.044418] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0094e3a8-99c0-4f90-869e-6421b1a94fd0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.059603] env[69171]: DEBUG nova.compute.provider_tree [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 780.132363] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5dc48c81-d7a6-4624-add3-fdd17cfaca47 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "6cdbe208-0eb2-4493-a477-daeb70739f95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.382s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.171135] env[69171]: DEBUG nova.network.neutron [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.589581] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Acquiring lock "776e2ab5-db02-46c8-a971-08b3c15e903f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.589776] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Lock "776e2ab5-db02-46c8-a971-08b3c15e903f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.590732] env[69171]: ERROR nova.scheduler.client.report [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [req-9357b75e-315d-45c3-80e0-0e69024c10d0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9357b75e-315d-45c3-80e0-0e69024c10d0"}]} [ 780.591090] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.155s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.591666] env[69171]: ERROR nova.compute.manager [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 780.591666] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Traceback (most recent call last): [ 780.591666] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 780.591666] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] yield [ 780.591666] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 780.591666] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] self.set_inventory_for_provider( [ 780.591666] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 780.591666] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9357b75e-315d-45c3-80e0-0e69024c10d0"}]} [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] During handling of the above exception, another exception occurred: [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Traceback (most recent call last): [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] with self.rt.instance_claim(context, instance, node, allocs, [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 780.591986] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] return f(*args, **kwargs) [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] self._update(elevated, cn) [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] self._update_to_placement(context, compute_node, startup) [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] return attempt.get(self._wrap_exception) [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] six.reraise(self.value[0], self.value[1], self.value[2]) [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] raise value [ 780.592337] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] self.reportclient.update_from_provider_tree( [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] with catch_all(pd.uuid): [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] self.gen.throw(typ, value, traceback) [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] raise exception.ResourceProviderSyncFailed() [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 780.592846] env[69171]: ERROR nova.compute.manager [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] [ 780.593243] env[69171]: DEBUG nova.compute.utils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 780.593608] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.097s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.594994] env[69171]: INFO nova.compute.claims [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 780.599584] env[69171]: DEBUG nova.compute.manager [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Build of instance 9a29b0d3-9de0-476b-8559-672010f02caa was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 780.599584] env[69171]: DEBUG nova.compute.manager [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 780.599584] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "refresh_cache-9a29b0d3-9de0-476b-8559-672010f02caa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.599795] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquired lock "refresh_cache-9a29b0d3-9de0-476b-8559-672010f02caa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.603017] env[69171]: DEBUG nova.network.neutron [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 780.637803] env[69171]: DEBUG nova.compute.manager [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 780.675925] env[69171]: INFO nova.compute.manager [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] [instance: 9bd1c816-5248-4e48-89f8-f3b999d50bff] Took 1.06 seconds to deallocate network for instance. [ 780.728820] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "f13b2145-5f8b-4b0a-ac07-c50018082fe2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.729239] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "f13b2145-5f8b-4b0a-ac07-c50018082fe2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.126257] env[69171]: DEBUG nova.network.neutron [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 781.174130] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 781.350656] env[69171]: DEBUG nova.network.neutron [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.637234] env[69171]: DEBUG nova.scheduler.client.report [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 781.656803] env[69171]: DEBUG nova.scheduler.client.report [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 781.657140] env[69171]: DEBUG nova.compute.provider_tree [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 781.676641] env[69171]: DEBUG nova.scheduler.client.report [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 781.713276] env[69171]: DEBUG nova.scheduler.client.report [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 781.728656] env[69171]: INFO nova.scheduler.client.report [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Deleted allocations for instance 9bd1c816-5248-4e48-89f8-f3b999d50bff [ 781.853621] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Releasing lock "refresh_cache-9a29b0d3-9de0-476b-8559-672010f02caa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.853884] env[69171]: DEBUG nova.compute.manager [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 781.854430] env[69171]: DEBUG nova.compute.manager [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 781.856068] env[69171]: DEBUG nova.network.neutron [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 781.873154] env[69171]: DEBUG nova.network.neutron [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 782.045432] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f57e80-ec4c-4ae1-9a36-6c31e2235b3d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.053934] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a784e95d-eb28-40e9-a102-6953d36d93ba {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.084988] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936a8c6f-276e-40dc-923d-683984e4892f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.093060] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-313c1698-0646-4163-9222-96bb77734146 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.107009] env[69171]: DEBUG nova.compute.provider_tree [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 782.239642] env[69171]: DEBUG oslo_concurrency.lockutils [None req-72c93de0-4cae-4d38-8f95-213e127e3bc1 tempest-ImagesTestJSON-1786329869 tempest-ImagesTestJSON-1786329869-project-member] Lock "9bd1c816-5248-4e48-89f8-f3b999d50bff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.273s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.377473] env[69171]: DEBUG nova.network.neutron [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 782.637293] env[69171]: ERROR nova.scheduler.client.report [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [req-8c87ff8a-7e29-44ad-88cf-d0fc9c291bc4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8c87ff8a-7e29-44ad-88cf-d0fc9c291bc4"}]} [ 782.637642] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.044s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.638275] env[69171]: ERROR nova.compute.manager [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 782.638275] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Traceback (most recent call last): [ 782.638275] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 782.638275] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] yield [ 782.638275] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 782.638275] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] self.set_inventory_for_provider( [ 782.638275] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 782.638275] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8c87ff8a-7e29-44ad-88cf-d0fc9c291bc4"}]} [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] During handling of the above exception, another exception occurred: [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Traceback (most recent call last): [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] with self.rt.instance_claim(context, instance, node, allocs, [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 782.638722] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] return f(*args, **kwargs) [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] self._update(elevated, cn) [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] self._update_to_placement(context, compute_node, startup) [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] return attempt.get(self._wrap_exception) [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] raise value [ 782.639070] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] self.reportclient.update_from_provider_tree( [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] with catch_all(pd.uuid): [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] self.gen.throw(typ, value, traceback) [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] raise exception.ResourceProviderSyncFailed() [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 782.639419] env[69171]: ERROR nova.compute.manager [instance: 221be95e-aac8-4551-8525-9745480ea8c6] [ 782.639731] env[69171]: DEBUG nova.compute.utils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 782.642903] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 18.108s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.643120] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.643293] env[69171]: INFO nova.compute.manager [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Successfully reverted task state from None on failure for instance. [ 782.646037] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.934s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 782.648336] env[69171]: INFO nova.compute.claims [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 782.650337] env[69171]: DEBUG nova.compute.manager [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Build of instance 221be95e-aac8-4551-8525-9745480ea8c6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 782.650773] env[69171]: DEBUG nova.compute.manager [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 782.651218] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Acquiring lock "refresh_cache-221be95e-aac8-4551-8525-9745480ea8c6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 782.651284] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Acquired lock "refresh_cache-221be95e-aac8-4551-8525-9745480ea8c6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.652375] env[69171]: DEBUG nova.network.neutron [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server [None req-a6b61ffc-8cc3-4c60-acf0-f168659cee63 tempest-ServersAdminNegativeTestJSON-825287255 tempest-ServersAdminNegativeTestJSON-825287255-project-member] Exception during message handling: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server yield [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server self.set_inventory_for_provider( [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6d1db241-64b5-4d4f-bfcc-321c7eca36e0"}]} [ 782.653201] env[69171]: ERROR oslo_messaging.rpc.server [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server During handling of the above exception, another exception occurred: [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server Traceback (most recent call last): [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/server.py", line 172, in _process_incoming [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server res = self.dispatcher.dispatch(message) [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 309, in dispatch [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server return self._do_dispatch(endpoint, method, ctxt, args) [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_messaging/rpc/dispatcher.py", line 229, in _do_dispatch [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server result = func(ctxt, **new_args) [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 65, in wrapped [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/exception_wrapper.py", line 63, in wrapped [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server return f(self, context, *args, **kw) [ 782.653541] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 166, in decorated_function [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 157, in decorated_function [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/utils.py", line 1453, in decorated_function [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 213, in decorated_function [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 203, in decorated_function [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server return function(self, context, *args, **kwargs) [ 782.654034] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3354, in terminate_instance [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server do_terminate_instance(instance, bdms) [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3349, in do_terminate_instance [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server with excutils.save_and_reraise_exception(): [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server self.force_reraise() [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server raise self.value [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3342, in do_terminate_instance [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server self._delete_instance(context, instance, bdms) [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 3306, in _delete_instance [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server self._complete_deletion(context, instance) [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 927, in _complete_deletion [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server self._update_resource_tracker(context, instance) [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/manager.py", line 693, in _update_resource_tracker [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server self.rt.update_usage(context, instance, instance.node) [ 782.654526] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server return f(*args, **kwargs) [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 702, in update_usage [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server self._update(context.elevated(), self.compute_nodes[nodename]) [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server self._update_to_placement(context, compute_node, startup) [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server return attempt.get(self._wrap_exception) [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server six.reraise(self.value[0], self.value[1], self.value[2]) [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server raise value [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 782.654972] env[69171]: ERROR oslo_messaging.rpc.server self.reportclient.update_from_provider_tree( [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server with catch_all(pd.uuid): [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server self.gen.throw(typ, value, traceback) [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server raise exception.ResourceProviderSyncFailed() [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 782.655458] env[69171]: ERROR oslo_messaging.rpc.server [ 782.744159] env[69171]: DEBUG nova.compute.manager [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 782.882767] env[69171]: INFO nova.compute.manager [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 9a29b0d3-9de0-476b-8559-672010f02caa] Took 1.03 seconds to deallocate network for instance. [ 783.181660] env[69171]: DEBUG nova.network.neutron [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.273140] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.346381] env[69171]: DEBUG nova.network.neutron [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.572161] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "f0af7537-ef12-4a21-8c4a-8afc46be9e0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 783.572425] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "f0af7537-ef12-4a21-8c4a-8afc46be9e0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.691595] env[69171]: DEBUG nova.scheduler.client.report [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 783.717981] env[69171]: DEBUG nova.scheduler.client.report [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 783.717981] env[69171]: DEBUG nova.compute.provider_tree [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 783.737602] env[69171]: DEBUG nova.scheduler.client.report [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 783.763374] env[69171]: DEBUG nova.scheduler.client.report [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 783.854239] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Releasing lock "refresh_cache-221be95e-aac8-4551-8525-9745480ea8c6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.854239] env[69171]: DEBUG nova.compute.manager [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 783.854239] env[69171]: DEBUG nova.compute.manager [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 783.854239] env[69171]: DEBUG nova.network.neutron [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 783.873745] env[69171]: DEBUG nova.network.neutron [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 783.916403] env[69171]: INFO nova.scheduler.client.report [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Deleted allocations for instance 9a29b0d3-9de0-476b-8559-672010f02caa [ 784.114290] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fd4ce4-f6e7-4b14-be26-085563fae109 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.122803] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3513cea-752a-4676-a2e6-dc5fc261daee {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.158620] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4ea200d-bf8d-4005-9625-4e6f24f3b15f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.167296] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36692195-df4c-4671-98c9-19b74af22cac {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.189130] env[69171]: DEBUG nova.compute.provider_tree [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 784.376436] env[69171]: DEBUG nova.network.neutron [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.435221] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f8130389-8331-4232-8591-8c09f76e90eb tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "9a29b0d3-9de0-476b-8559-672010f02caa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.270s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.727452] env[69171]: ERROR nova.scheduler.client.report [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [req-8b7b51d0-15a7-4b01-b065-0314e15c06d2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8b7b51d0-15a7-4b01-b065-0314e15c06d2"}]} [ 784.730078] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.079s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.730078] env[69171]: ERROR nova.compute.manager [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 784.730078] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Traceback (most recent call last): [ 784.730078] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 784.730078] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] yield [ 784.730078] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 784.730078] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] self.set_inventory_for_provider( [ 784.730078] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 784.730078] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8b7b51d0-15a7-4b01-b065-0314e15c06d2"}]} [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] During handling of the above exception, another exception occurred: [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Traceback (most recent call last): [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] with self.rt.instance_claim(context, instance, node, allocs, [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 784.730397] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] return f(*args, **kwargs) [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] self._update(elevated, cn) [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] self._update_to_placement(context, compute_node, startup) [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] return attempt.get(self._wrap_exception) [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] six.reraise(self.value[0], self.value[1], self.value[2]) [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] raise value [ 784.730702] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] self.reportclient.update_from_provider_tree( [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] with catch_all(pd.uuid): [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] self.gen.throw(typ, value, traceback) [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] raise exception.ResourceProviderSyncFailed() [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 784.731100] env[69171]: ERROR nova.compute.manager [instance: 47c2201c-a952-4474-88e1-2773f04494aa] [ 784.731374] env[69171]: DEBUG nova.compute.utils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 784.732231] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.487s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.739757] env[69171]: DEBUG nova.compute.manager [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Build of instance 47c2201c-a952-4474-88e1-2773f04494aa was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 784.739757] env[69171]: DEBUG nova.compute.manager [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 784.739757] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Acquiring lock "refresh_cache-47c2201c-a952-4474-88e1-2773f04494aa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.739757] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Acquired lock "refresh_cache-47c2201c-a952-4474-88e1-2773f04494aa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.740016] env[69171]: DEBUG nova.network.neutron [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 784.885759] env[69171]: INFO nova.compute.manager [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] [instance: 221be95e-aac8-4551-8525-9745480ea8c6] Took 1.03 seconds to deallocate network for instance. [ 784.937623] env[69171]: DEBUG nova.compute.manager [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 785.198456] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquiring lock "aed06808-707f-4d80-ad42-bdd2e8f90cd6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.199710] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "aed06808-707f-4d80-ad42-bdd2e8f90cd6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.288396] env[69171]: DEBUG nova.network.neutron [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 785.400557] env[69171]: DEBUG nova.network.neutron [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.468040] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.750147] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Applying migration context for instance f8b20421-15d3-46f3-a504-172d044ff4d3 as it has an incoming, in-progress migration 7bd57b7a-be81-4c69-92f8-50364ba09d44. Migration status is error {{(pid=69171) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 785.909267] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Releasing lock "refresh_cache-47c2201c-a952-4474-88e1-2773f04494aa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.909500] env[69171]: DEBUG nova.compute.manager [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 785.909694] env[69171]: DEBUG nova.compute.manager [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 785.909862] env[69171]: DEBUG nova.network.neutron [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 785.940276] env[69171]: INFO nova.scheduler.client.report [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Deleted allocations for instance 221be95e-aac8-4551-8525-9745480ea8c6 [ 785.947771] env[69171]: DEBUG nova.network.neutron [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 786.282272] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "211d6fa6-6d34-45e8-bfc3-b5feccb642de" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.282272] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "211d6fa6-6d34-45e8-bfc3-b5feccb642de" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.455085] env[69171]: DEBUG nova.network.neutron [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.458539] env[69171]: DEBUG oslo_concurrency.lockutils [None req-53245a42-e67b-43ed-83b6-60a4972b2799 tempest-FloatingIPsAssociationTestJSON-537293438 tempest-FloatingIPsAssociationTestJSON-537293438-project-member] Lock "221be95e-aac8-4551-8525-9745480ea8c6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.522s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.958328] env[69171]: INFO nova.compute.manager [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] [instance: 47c2201c-a952-4474-88e1-2773f04494aa] Took 1.05 seconds to deallocate network for instance. [ 786.960988] env[69171]: DEBUG nova.compute.manager [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 787.496267] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 787.767342] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Skipping migration as instance is neither resizing nor live-migrating. {{(pid=69171) _update_usage_from_migrations /opt/stack/nova/nova/compute/resource_tracker.py:1563}} [ 787.790439] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f5a71cca-1e32-4c3c-9bb8-83e634a864f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.790638] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 8e6a9d73-6705-484c-82e0-70c6a6860211 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.790789] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 6a29d495-c938-41ad-97b6-1331b6eff589 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.790936] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 787.791156] env[69171]: WARNING nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f8b20421-15d3-46f3-a504-172d044ff4d3 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 787.791313] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 47c2201c-a952-4474-88e1-2773f04494aa actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 788.008561] env[69171]: INFO nova.scheduler.client.report [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Deleted allocations for instance 47c2201c-a952-4474-88e1-2773f04494aa [ 788.295340] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 788.516519] env[69171]: DEBUG oslo_concurrency.lockutils [None req-c0013b0b-c73d-43a9-8942-b8a34dcd3b55 tempest-ServersNegativeTestMultiTenantJSON-1483829143 tempest-ServersNegativeTestMultiTenantJSON-1483829143-project-member] Lock "47c2201c-a952-4474-88e1-2773f04494aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.995s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 788.800513] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 05eacaa7-1a14-4834-86ee-7f6dcaf21b85 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.020633] env[69171]: DEBUG nova.compute.manager [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 789.304654] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 59600940-1b15-4192-906d-936d15b4377c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 789.556104] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 789.809974] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.315021] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 62fecdeb-6657-4fc5-934c-5c459e7287d7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 790.818022] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 6664e9ec-4c43-464d-9117-92967031fe55 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.322196] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 791.446451] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquiring lock "c0976d99-04aa-4685-8408-ff5e69de129f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.446451] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "c0976d99-04aa-4685-8408-ff5e69de129f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.663552] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "b189c5ed-d475-44ec-b26e-e7c7014f0a81" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.664009] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "b189c5ed-d475-44ec-b26e-e7c7014f0a81" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.826011] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.331405] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 54d5f178-7fc5-48ce-be89-c783f419e581 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 792.450915] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquiring lock "c5179486-ea0d-4664-8790-9ec0b9eab2ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.451486] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "c5179486-ea0d-4664-8790-9ec0b9eab2ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 792.836962] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 478e1a26-101c-48a4-bcc2-a103f59962f6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.093592] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "36e1bcc1-8503-4786-b3aa-381a110e8bf6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.093826] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "36e1bcc1-8503-4786-b3aa-381a110e8bf6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.342842] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance ed037ce0-072f-425a-97fa-5fb0f941268b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 793.697302] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "8936c08f-2ef1-4b6e-a632-60ca9ad7095d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.697557] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "8936c08f-2ef1-4b6e-a632-60ca9ad7095d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.846454] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 776e2ab5-db02-46c8-a971-08b3c15e903f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.189447] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Acquiring lock "2d5a9f39-c7c1-4aab-91c0-54b51547eab4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.189447] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Lock "2d5a9f39-c7c1-4aab-91c0-54b51547eab4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 794.349864] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f13b2145-5f8b-4b0a-ac07-c50018082fe2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 794.857027] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f0af7537-ef12-4a21-8c4a-8afc46be9e0b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.357431] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance aed06808-707f-4d80-ad42-bdd2e8f90cd6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.860800] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 211d6fa6-6d34-45e8-bfc3-b5feccb642de has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 795.861291] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 795.861872] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 795.890210] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 795.906429] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 795.906629] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 795.925074] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 795.960590] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 796.417540] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c91d47-e323-4e81-9bc7-95933a6c1c75 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.427470] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a791565a-17bc-468f-b607-751a8d299972 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.461024] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efa90433-8263-4400-8e46-88c017a1af21 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.466879] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67c96c96-547c-4628-af47-7e770f46f74d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.486306] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 797.012343] env[69171]: ERROR nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [req-8d8a079c-3800-4adb-b2a3-95b02d79c86b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8d8a079c-3800-4adb-b2a3-95b02d79c86b"}]} [ 797.012692] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 12.281s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.013715] env[69171]: ERROR nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Error updating resources for node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28.: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 797.013715] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 797.013715] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 797.013715] env[69171]: ERROR nova.compute.manager yield [ 797.013715] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 797.013715] env[69171]: ERROR nova.compute.manager self.set_inventory_for_provider( [ 797.013715] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 797.013715] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 797.013715] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8d8a079c-3800-4adb-b2a3-95b02d79c86b"}]} [ 797.013715] env[69171]: ERROR nova.compute.manager [ 797.013715] env[69171]: ERROR nova.compute.manager During handling of the above exception, another exception occurred: [ 797.013715] env[69171]: ERROR nova.compute.manager [ 797.014186] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 10584, in _update_available_resource_for_node [ 797.014186] env[69171]: ERROR nova.compute.manager self.rt.update_available_resource(context, nodename, [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 935, in update_available_resource [ 797.014186] env[69171]: ERROR nova.compute.manager self._update_available_resource(context, resources, startup=startup) [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 797.014186] env[69171]: ERROR nova.compute.manager return f(*args, **kwargs) [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1066, in _update_available_resource [ 797.014186] env[69171]: ERROR nova.compute.manager self._update(context, cn, startup=startup) [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 797.014186] env[69171]: ERROR nova.compute.manager self._update_to_placement(context, compute_node, startup) [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 797.014186] env[69171]: ERROR nova.compute.manager return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 797.014186] env[69171]: ERROR nova.compute.manager return attempt.get(self._wrap_exception) [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 797.014186] env[69171]: ERROR nova.compute.manager six.reraise(self.value[0], self.value[1], self.value[2]) [ 797.014186] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 797.014884] env[69171]: ERROR nova.compute.manager raise value [ 797.014884] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 797.014884] env[69171]: ERROR nova.compute.manager attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 797.014884] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 797.014884] env[69171]: ERROR nova.compute.manager self.reportclient.update_from_provider_tree( [ 797.014884] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 797.014884] env[69171]: ERROR nova.compute.manager with catch_all(pd.uuid): [ 797.014884] env[69171]: ERROR nova.compute.manager File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 797.014884] env[69171]: ERROR nova.compute.manager self.gen.throw(typ, value, traceback) [ 797.014884] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 797.014884] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderSyncFailed() [ 797.014884] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 797.014884] env[69171]: ERROR nova.compute.manager [ 797.014884] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.412s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.016291] env[69171]: INFO nova.compute.claims [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 797.876369] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Acquiring lock "8e90508a-fc26-4e9d-ac70-054b249bc2a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 797.876463] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Lock "8e90508a-fc26-4e9d-ac70-054b249bc2a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.045617] env[69171]: DEBUG nova.scheduler.client.report [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 798.065741] env[69171]: DEBUG nova.scheduler.client.report [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 798.065992] env[69171]: DEBUG nova.compute.provider_tree [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 798.085971] env[69171]: DEBUG nova.scheduler.client.report [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 798.103481] env[69171]: DEBUG nova.scheduler.client.report [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 798.570714] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70407f1-4e7f-4b87-b506-6c6a533198e1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.581726] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a750a703-5487-40e2-8f09-c22a0741eccb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.616296] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4410652-163f-415d-bc2b-ba1df6e7474f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.627035] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2455cc4-338e-4648-b746-0822aafb3004 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.642352] env[69171]: DEBUG nova.compute.provider_tree [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 799.164391] env[69171]: ERROR nova.scheduler.client.report [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [req-1d4c1a3d-62b9-4f06-a83a-9616bfc3d888] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1d4c1a3d-62b9-4f06-a83a-9616bfc3d888"}]} [ 799.164789] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.151s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 799.165833] env[69171]: ERROR nova.compute.manager [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 799.165833] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Traceback (most recent call last): [ 799.165833] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 799.165833] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] yield [ 799.165833] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 799.165833] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] self.set_inventory_for_provider( [ 799.165833] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 799.165833] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1d4c1a3d-62b9-4f06-a83a-9616bfc3d888"}]} [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] During handling of the above exception, another exception occurred: [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Traceback (most recent call last): [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] with self.rt.instance_claim(context, instance, node, allocs, [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 799.166215] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] return f(*args, **kwargs) [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] self._update(elevated, cn) [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] self._update_to_placement(context, compute_node, startup) [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] return attempt.get(self._wrap_exception) [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] six.reraise(self.value[0], self.value[1], self.value[2]) [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] raise value [ 799.166579] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] self.reportclient.update_from_provider_tree( [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] with catch_all(pd.uuid): [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] self.gen.throw(typ, value, traceback) [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] raise exception.ResourceProviderSyncFailed() [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 799.167105] env[69171]: ERROR nova.compute.manager [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] [ 799.167458] env[69171]: DEBUG nova.compute.utils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 799.169996] env[69171]: DEBUG nova.compute.manager [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Build of instance a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 799.170261] env[69171]: DEBUG nova.compute.manager [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 799.170517] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "refresh_cache-a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.170663] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquired lock "refresh_cache-a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.170819] env[69171]: DEBUG nova.network.neutron [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 799.172706] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.963s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 799.174392] env[69171]: INFO nova.compute.claims [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 799.702102] env[69171]: DEBUG nova.network.neutron [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 799.785261] env[69171]: DEBUG nova.network.neutron [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.206243] env[69171]: DEBUG nova.scheduler.client.report [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 800.234155] env[69171]: DEBUG nova.scheduler.client.report [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 800.234434] env[69171]: DEBUG nova.compute.provider_tree [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 800.248841] env[69171]: DEBUG nova.scheduler.client.report [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 800.270916] env[69171]: DEBUG nova.scheduler.client.report [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 800.287474] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Releasing lock "refresh_cache-a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.287712] env[69171]: DEBUG nova.compute.manager [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 800.287895] env[69171]: DEBUG nova.compute.manager [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 800.288073] env[69171]: DEBUG nova.network.neutron [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 800.308100] env[69171]: DEBUG nova.network.neutron [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 800.453526] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquiring lock "a0941eb1-4a98-450b-8318-e8f60db1ac03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.453769] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "a0941eb1-4a98-450b-8318-e8f60db1ac03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.678043] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5420c874-a822-4f72-bd76-714710cb269d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.686399] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02a8a3b5-4147-4158-a73c-e70822f06957 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.717294] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2f7d6a-984e-4dca-8963-3264206326fb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.725839] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1826814e-f318-456b-8ced-986a74a727cc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.740474] env[69171]: DEBUG nova.compute.provider_tree [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 800.821180] env[69171]: DEBUG nova.network.neutron [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.261280] env[69171]: ERROR nova.scheduler.client.report [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [req-b43e128c-5b5b-4b5a-ac51-5c9b096dab4d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b43e128c-5b5b-4b5a-ac51-5c9b096dab4d"}]} [ 801.261685] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.089s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 801.262318] env[69171]: ERROR nova.compute.manager [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 801.262318] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Traceback (most recent call last): [ 801.262318] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 801.262318] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] yield [ 801.262318] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 801.262318] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] self.set_inventory_for_provider( [ 801.262318] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 801.262318] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b43e128c-5b5b-4b5a-ac51-5c9b096dab4d"}]} [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] During handling of the above exception, another exception occurred: [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Traceback (most recent call last): [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] with self.rt.instance_claim(context, instance, node, allocs, [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 801.262613] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] return f(*args, **kwargs) [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] self._update(elevated, cn) [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] self._update_to_placement(context, compute_node, startup) [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] return attempt.get(self._wrap_exception) [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] six.reraise(self.value[0], self.value[1], self.value[2]) [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] raise value [ 801.262928] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] self.reportclient.update_from_provider_tree( [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] with catch_all(pd.uuid): [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] self.gen.throw(typ, value, traceback) [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] raise exception.ResourceProviderSyncFailed() [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 801.263412] env[69171]: ERROR nova.compute.manager [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] [ 801.263777] env[69171]: DEBUG nova.compute.utils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 801.264315] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.309s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 801.266114] env[69171]: INFO nova.compute.claims [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 801.268854] env[69171]: DEBUG nova.compute.manager [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Build of instance 05eacaa7-1a14-4834-86ee-7f6dcaf21b85 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 801.269319] env[69171]: DEBUG nova.compute.manager [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 801.269556] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquiring lock "refresh_cache-05eacaa7-1a14-4834-86ee-7f6dcaf21b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.269742] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Acquired lock "refresh_cache-05eacaa7-1a14-4834-86ee-7f6dcaf21b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.269913] env[69171]: DEBUG nova.network.neutron [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 801.324026] env[69171]: INFO nova.compute.manager [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8] Took 1.04 seconds to deallocate network for instance. [ 801.790956] env[69171]: DEBUG nova.network.neutron [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 801.871752] env[69171]: DEBUG nova.network.neutron [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.296729] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 802.310771] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 802.311198] env[69171]: DEBUG nova.compute.provider_tree [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 802.324079] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 802.343020] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 802.354496] env[69171]: INFO nova.scheduler.client.report [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Deleted allocations for instance a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8 [ 802.374835] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Releasing lock "refresh_cache-05eacaa7-1a14-4834-86ee-7f6dcaf21b85" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.374984] env[69171]: DEBUG nova.compute.manager [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 802.375840] env[69171]: DEBUG nova.compute.manager [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 802.375840] env[69171]: DEBUG nova.network.neutron [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 802.392873] env[69171]: DEBUG nova.network.neutron [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 802.722910] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0ca87e3-8c1e-4874-bb9e-3c7f75203cdc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.733510] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9669c47a-d2c5-4730-8d25-fdb6d2f0d219 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.782411] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a0efbd6-8394-4f90-a483-724573d21dd7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.794450] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7f55309-4d57-4c17-953d-ec9b352e1a2c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.817924] env[69171]: DEBUG nova.compute.provider_tree [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 802.871025] env[69171]: DEBUG oslo_concurrency.lockutils [None req-78e5f0ac-e52a-4945-9b5d-e85257e0fe8b tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "a1a8aca5-e4b3-4927-9c7d-b86c58c9c9e8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.266s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.897239] env[69171]: DEBUG nova.network.neutron [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.339411] env[69171]: ERROR nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [req-d72c13b2-417a-43a0-a393-6cec5bc61992] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d72c13b2-417a-43a0-a393-6cec5bc61992"}]} [ 803.339775] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.075s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.340759] env[69171]: ERROR nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 803.340759] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] Traceback (most recent call last): [ 803.340759] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 803.340759] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] yield [ 803.340759] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 803.340759] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] self.set_inventory_for_provider( [ 803.340759] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 803.340759] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d72c13b2-417a-43a0-a393-6cec5bc61992"}]} [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] During handling of the above exception, another exception occurred: [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] Traceback (most recent call last): [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] with self.rt.instance_claim(context, instance, node, allocs, [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 803.341048] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] return f(*args, **kwargs) [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] self._update(elevated, cn) [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] self._update_to_placement(context, compute_node, startup) [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] return attempt.get(self._wrap_exception) [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] six.reraise(self.value[0], self.value[1], self.value[2]) [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] raise value [ 803.341395] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] self.reportclient.update_from_provider_tree( [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] with catch_all(pd.uuid): [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] self.gen.throw(typ, value, traceback) [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] raise exception.ResourceProviderSyncFailed() [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 803.341738] env[69171]: ERROR nova.compute.manager [instance: 59600940-1b15-4192-906d-936d15b4377c] [ 803.342043] env[69171]: DEBUG nova.compute.utils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 803.342828] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 29.323s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.343031] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.344962] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.567s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.346425] env[69171]: INFO nova.compute.claims [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 803.349157] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Build of instance 59600940-1b15-4192-906d-936d15b4377c was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 803.349593] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 803.349867] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "refresh_cache-59600940-1b15-4192-906d-936d15b4377c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.350056] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquired lock "refresh_cache-59600940-1b15-4192-906d-936d15b4377c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.350172] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 803.370456] env[69171]: INFO nova.scheduler.client.report [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Deleted allocations for instance f8b20421-15d3-46f3-a504-172d044ff4d3 [ 803.375802] env[69171]: DEBUG nova.compute.manager [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 803.400062] env[69171]: INFO nova.compute.manager [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] [instance: 05eacaa7-1a14-4834-86ee-7f6dcaf21b85] Took 1.02 seconds to deallocate network for instance. [ 803.878124] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 803.884908] env[69171]: DEBUG oslo_concurrency.lockutils [None req-60f7d47b-f501-4753-99a9-af7773829166 tempest-MigrationsAdminTest-2031345157 tempest-MigrationsAdminTest-2031345157-project-member] Lock "f8b20421-15d3-46f3-a504-172d044ff4d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 32.421s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.886572] env[69171]: DEBUG oslo_concurrency.lockutils [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] Acquired lock "f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.889830] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-040c7261-2558-449b-947c-5a5222028fe5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.912509] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 803.913695] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb836c5-5aaf-4f91-9fd5-168b1e618c0a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.940064] env[69171]: DEBUG nova.virt.vmwareapi.vmops [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Reconfiguring VM to detach interface {{(pid=69171) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1974}} [ 803.943540] env[69171]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bdd4c10-9f7b-43f4-b467-3020649f9a85 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.963673] env[69171]: DEBUG oslo_vmware.api [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] Waiting for the task: (returnval){ [ 803.963673] env[69171]: value = "task-4235951" [ 803.963673] env[69171]: _type = "Task" [ 803.963673] env[69171]: } to complete. {{(pid=69171) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.975233] env[69171]: DEBUG oslo_vmware.api [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] Task: {'id': task-4235951, 'name': ReconfigVM_Task} progress is 6%. {{(pid=69171) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.993757] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.381038] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 804.395892] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 804.396178] env[69171]: DEBUG nova.compute.provider_tree [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.408253] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 804.427505] env[69171]: DEBUG nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 804.439580] env[69171]: INFO nova.scheduler.client.report [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Deleted allocations for instance 05eacaa7-1a14-4834-86ee-7f6dcaf21b85 [ 804.475063] env[69171]: DEBUG oslo_vmware.exceptions [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] Fault QuestionPending not matched. {{(pid=69171) get_fault_class /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/exceptions.py:290}} [ 804.479051] env[69171]: ERROR nova.virt.vmwareapi.vmops [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Detaching network adapter failed. Exception: The operation cannot be allowed at the current time because the virtual machine has a question pending: [ 804.479051] env[69171]: 'msg.hbacommon.outofspace:There is no more space for virtual disk 'f8b20421-15d3-46f3-a504-172d044ff4d3.vmdk'. You might be able to continue this session by freeing disk space on the relevant volume, and clicking Retry. Click Cancel to terminate this session. [ 804.479051] env[69171]: '. [ 804.479051] env[69171]: Faults: ['QuestionPending']: oslo_vmware.exceptions.VimFaultException: The operation cannot be allowed at the current time because the virtual machine has a question pending: [ 804.479281] env[69171]: DEBUG oslo_concurrency.lockutils [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] Releasing lock "f8b20421-15d3-46f3-a504-172d044ff4d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.479613] env[69171]: WARNING nova.compute.manager [req-cb23f6b8-e610-42df-abcb-2292b5beaeda req-0da54a09-b844-4dfa-89d2-33b060f5319b service nova] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Detach interface failed, port_id=8cd0dd84-cea0-4ec8-a140-1f2b282ab7e7, reason: Failed to detach network adapter device from f8b20421-15d3-46f3-a504-172d044ff4d3: nova.exception.InterfaceDetachFailed: Failed to detach network adapter device from f8b20421-15d3-46f3-a504-172d044ff4d3 [ 804.483048] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "ecec698a-7570-4400-ad8c-9215cd6733ef" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 804.483263] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "ecec698a-7570-4400-ad8c-9215cd6733ef" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 804.496741] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Releasing lock "refresh_cache-59600940-1b15-4192-906d-936d15b4377c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.497103] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 804.497391] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 804.497580] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 804.515933] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 804.795617] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19d736e4-73fc-4e47-82f8-1fa38ecaaf37 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.804275] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ff9b1e6-6713-4c93-a0c0-d5ecffb5041b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.834117] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34b1e64-8982-4ee0-86c2-0e81b0e46adc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.842429] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d1a721-82b4-4030-ac9b-75c1f0744977 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.857718] env[69171]: DEBUG nova.compute.provider_tree [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 804.950299] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8bc26413-0271-4e1d-9998-64b7f8b8daab tempest-ServersTestMultiNic-1378921423 tempest-ServersTestMultiNic-1378921423-project-member] Lock "05eacaa7-1a14-4834-86ee-7f6dcaf21b85" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.722s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.019184] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.384138] env[69171]: ERROR nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [req-0ef24144-23e1-43f4-8326-9dbb271fa4dd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0ef24144-23e1-43f4-8326-9dbb271fa4dd"}]} [ 805.384138] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.037s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.384534] env[69171]: ERROR nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 805.384534] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Traceback (most recent call last): [ 805.384534] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 805.384534] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] yield [ 805.384534] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 805.384534] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] self.set_inventory_for_provider( [ 805.384534] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 805.384534] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0ef24144-23e1-43f4-8326-9dbb271fa4dd"}]} [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] During handling of the above exception, another exception occurred: [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Traceback (most recent call last): [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] with self.rt.instance_claim(context, instance, node, allocs, [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 805.384778] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] return f(*args, **kwargs) [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] self._update(elevated, cn) [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] self._update_to_placement(context, compute_node, startup) [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] return attempt.get(self._wrap_exception) [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] six.reraise(self.value[0], self.value[1], self.value[2]) [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] raise value [ 805.385079] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] self.reportclient.update_from_provider_tree( [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] with catch_all(pd.uuid): [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] self.gen.throw(typ, value, traceback) [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] raise exception.ResourceProviderSyncFailed() [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 805.385475] env[69171]: ERROR nova.compute.manager [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] [ 805.385786] env[69171]: DEBUG nova.compute.utils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 805.386122] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.540s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.387712] env[69171]: INFO nova.compute.claims [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 805.390671] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Build of instance 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 805.391256] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 805.391606] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquiring lock "refresh_cache-55cf3ad8-d03c-4ac0-9c48-0a9382bc8628" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.391861] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Acquired lock "refresh_cache-55cf3ad8-d03c-4ac0-9c48-0a9382bc8628" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.392171] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 805.454543] env[69171]: DEBUG nova.compute.manager [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 805.521864] env[69171]: INFO nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 59600940-1b15-4192-906d-936d15b4377c] Took 1.02 seconds to deallocate network for instance. [ 805.917130] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 805.977202] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.055041] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.425778] env[69171]: DEBUG nova.scheduler.client.report [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 806.439678] env[69171]: DEBUG nova.scheduler.client.report [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 806.439945] env[69171]: DEBUG nova.compute.provider_tree [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 806.454962] env[69171]: DEBUG nova.scheduler.client.report [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 806.472779] env[69171]: DEBUG nova.scheduler.client.report [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 806.555881] env[69171]: INFO nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Deleted allocations for instance 59600940-1b15-4192-906d-936d15b4377c [ 806.562121] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Releasing lock "refresh_cache-55cf3ad8-d03c-4ac0-9c48-0a9382bc8628" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.562382] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 806.562581] env[69171]: DEBUG nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 806.562724] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 806.589842] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 806.879538] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce6764b-9c30-470f-b665-9dd63449dfba {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.887965] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9995b6b9-bb61-448c-ba87-c1db773b9727 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.920240] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32deceb-b614-456d-ae86-d6bd294e4c9f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.934756] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b91b93-8178-4d7c-ac77-ff13db29596d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.945172] env[69171]: DEBUG nova.compute.provider_tree [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 807.069334] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "59600940-1b15-4192-906d-936d15b4377c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.762s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.092925] env[69171]: DEBUG nova.network.neutron [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.466959] env[69171]: ERROR nova.scheduler.client.report [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [req-5b86de37-d7a7-4975-a000-12a37dfeb2df] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5b86de37-d7a7-4975-a000-12a37dfeb2df"}]} [ 807.467459] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.081s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.468018] env[69171]: ERROR nova.compute.manager [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 807.468018] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Traceback (most recent call last): [ 807.468018] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 807.468018] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] yield [ 807.468018] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 807.468018] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] self.set_inventory_for_provider( [ 807.468018] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 807.468018] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5b86de37-d7a7-4975-a000-12a37dfeb2df"}]} [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] During handling of the above exception, another exception occurred: [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Traceback (most recent call last): [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] with self.rt.instance_claim(context, instance, node, allocs, [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 807.468317] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] return f(*args, **kwargs) [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] self._update(elevated, cn) [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] self._update_to_placement(context, compute_node, startup) [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] return attempt.get(self._wrap_exception) [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] six.reraise(self.value[0], self.value[1], self.value[2]) [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] raise value [ 807.468681] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] self.reportclient.update_from_provider_tree( [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] with catch_all(pd.uuid): [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] self.gen.throw(typ, value, traceback) [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] raise exception.ResourceProviderSyncFailed() [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 807.469158] env[69171]: ERROR nova.compute.manager [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] [ 807.469551] env[69171]: DEBUG nova.compute.utils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.470452] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.298s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.472192] env[69171]: INFO nova.compute.claims [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 807.474966] env[69171]: DEBUG nova.compute.manager [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Build of instance 62fecdeb-6657-4fc5-934c-5c459e7287d7 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 807.476163] env[69171]: DEBUG nova.compute.manager [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 807.476429] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Acquiring lock "refresh_cache-62fecdeb-6657-4fc5-934c-5c459e7287d7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.476587] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Acquired lock "refresh_cache-62fecdeb-6657-4fc5-934c-5c459e7287d7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.476752] env[69171]: DEBUG nova.network.neutron [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 807.573044] env[69171]: DEBUG nova.compute.manager [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 807.595776] env[69171]: INFO nova.compute.manager [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] [instance: 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628] Took 1.03 seconds to deallocate network for instance. [ 807.997580] env[69171]: DEBUG nova.network.neutron [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.079169] env[69171]: DEBUG nova.network.neutron [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.092762] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.424462] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Acquiring lock "a74e930a-9871-4aa6-8997-e5ef6042c9d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 808.424699] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Lock "a74e930a-9871-4aa6-8997-e5ef6042c9d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 808.502028] env[69171]: DEBUG nova.scheduler.client.report [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 808.514930] env[69171]: DEBUG nova.scheduler.client.report [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 808.515176] env[69171]: DEBUG nova.compute.provider_tree [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 808.526761] env[69171]: DEBUG nova.scheduler.client.report [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 808.544888] env[69171]: DEBUG nova.scheduler.client.report [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 808.583209] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Releasing lock "refresh_cache-62fecdeb-6657-4fc5-934c-5c459e7287d7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.583492] env[69171]: DEBUG nova.compute.manager [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 808.583653] env[69171]: DEBUG nova.compute.manager [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 808.583822] env[69171]: DEBUG nova.network.neutron [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 808.600971] env[69171]: DEBUG nova.network.neutron [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 808.625899] env[69171]: INFO nova.scheduler.client.report [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Deleted allocations for instance 55cf3ad8-d03c-4ac0-9c48-0a9382bc8628 [ 808.852724] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4878840d-f927-46e6-aaf8-37d5dadba538 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.860579] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12019d1f-c609-4794-afcf-5701d9a7a3e5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.891229] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1c781a3-f0bb-4010-b255-1c84534ac2af {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.899139] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc8d45ab-a354-400f-8f1e-299a45829d2c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.913085] env[69171]: DEBUG nova.compute.provider_tree [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 809.103760] env[69171]: DEBUG nova.network.neutron [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.136023] env[69171]: DEBUG oslo_concurrency.lockutils [None req-db1a4811-b06a-40ee-bc0d-6c4b8484560f tempest-MultipleCreateTestJSON-1657510539 tempest-MultipleCreateTestJSON-1657510539-project-member] Lock "55cf3ad8-d03c-4ac0-9c48-0a9382bc8628" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.771s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.442401] env[69171]: ERROR nova.scheduler.client.report [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [req-ffcede3f-2535-4d02-a2d2-853d267aa592] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ffcede3f-2535-4d02-a2d2-853d267aa592"}]} [ 809.442854] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.972s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.443695] env[69171]: ERROR nova.compute.manager [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 809.443695] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Traceback (most recent call last): [ 809.443695] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 809.443695] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] yield [ 809.443695] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 809.443695] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] self.set_inventory_for_provider( [ 809.443695] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 809.443695] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ffcede3f-2535-4d02-a2d2-853d267aa592"}]} [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] During handling of the above exception, another exception occurred: [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Traceback (most recent call last): [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] with self.rt.instance_claim(context, instance, node, allocs, [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 809.444017] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] return f(*args, **kwargs) [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] self._update(elevated, cn) [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] self._update_to_placement(context, compute_node, startup) [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] return attempt.get(self._wrap_exception) [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] six.reraise(self.value[0], self.value[1], self.value[2]) [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] raise value [ 809.444273] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] self.reportclient.update_from_provider_tree( [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] with catch_all(pd.uuid): [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] self.gen.throw(typ, value, traceback) [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] raise exception.ResourceProviderSyncFailed() [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 809.444639] env[69171]: ERROR nova.compute.manager [instance: 6664e9ec-4c43-464d-9117-92967031fe55] [ 809.444926] env[69171]: DEBUG nova.compute.utils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 809.446384] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.173s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.449049] env[69171]: INFO nova.compute.claims [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 809.452316] env[69171]: DEBUG nova.compute.manager [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Build of instance 6664e9ec-4c43-464d-9117-92967031fe55 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 809.452789] env[69171]: DEBUG nova.compute.manager [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 809.454466] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Acquiring lock "refresh_cache-6664e9ec-4c43-464d-9117-92967031fe55" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.454694] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Acquired lock "refresh_cache-6664e9ec-4c43-464d-9117-92967031fe55" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.454990] env[69171]: DEBUG nova.network.neutron [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 809.606629] env[69171]: INFO nova.compute.manager [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] [instance: 62fecdeb-6657-4fc5-934c-5c459e7287d7] Took 1.02 seconds to deallocate network for instance. [ 809.638552] env[69171]: DEBUG nova.compute.manager [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 809.979779] env[69171]: DEBUG nova.network.neutron [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.054762] env[69171]: DEBUG nova.network.neutron [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.173738] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 810.495401] env[69171]: DEBUG nova.scheduler.client.report [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 810.511604] env[69171]: DEBUG nova.scheduler.client.report [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 810.511829] env[69171]: DEBUG nova.compute.provider_tree [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 810.524729] env[69171]: DEBUG nova.scheduler.client.report [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 810.555494] env[69171]: DEBUG nova.scheduler.client.report [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 810.557888] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Releasing lock "refresh_cache-6664e9ec-4c43-464d-9117-92967031fe55" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.558123] env[69171]: DEBUG nova.compute.manager [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 810.558309] env[69171]: DEBUG nova.compute.manager [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 810.558475] env[69171]: DEBUG nova.network.neutron [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 810.578461] env[69171]: DEBUG nova.network.neutron [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 810.641978] env[69171]: INFO nova.scheduler.client.report [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Deleted allocations for instance 62fecdeb-6657-4fc5-934c-5c459e7287d7 [ 810.927857] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7f32f4-f2a7-44a6-8baf-8b88bfd7457d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.936010] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ea4b29-c85f-48df-8437-96c1e1d1ca37 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.966608] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06182fec-2021-46dc-ab91-a91bd3ce27a7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.974456] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e9953ba-81d4-446c-90bb-c346a7071647 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.990030] env[69171]: DEBUG nova.compute.provider_tree [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 811.084393] env[69171]: DEBUG nova.network.neutron [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.152672] env[69171]: DEBUG oslo_concurrency.lockutils [None req-7c0025b5-ae86-4892-8bda-991b1cc129dc tempest-ServerMetadataTestJSON-1366804275 tempest-ServerMetadataTestJSON-1366804275-project-member] Lock "62fecdeb-6657-4fc5-934c-5c459e7287d7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.204s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.516583] env[69171]: ERROR nova.scheduler.client.report [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [req-296e7cd3-69b0-4cbc-8183-3eb8c08aabe7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-296e7cd3-69b0-4cbc-8183-3eb8c08aabe7"}]} [ 811.516958] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.071s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.517567] env[69171]: ERROR nova.compute.manager [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 811.517567] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Traceback (most recent call last): [ 811.517567] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 811.517567] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] yield [ 811.517567] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 811.517567] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] self.set_inventory_for_provider( [ 811.517567] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 811.517567] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-296e7cd3-69b0-4cbc-8183-3eb8c08aabe7"}]} [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] During handling of the above exception, another exception occurred: [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Traceback (most recent call last): [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] with self.rt.instance_claim(context, instance, node, allocs, [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 811.517828] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] return f(*args, **kwargs) [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] self._update(elevated, cn) [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] self._update_to_placement(context, compute_node, startup) [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] return attempt.get(self._wrap_exception) [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] six.reraise(self.value[0], self.value[1], self.value[2]) [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] raise value [ 811.518130] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] self.reportclient.update_from_provider_tree( [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] with catch_all(pd.uuid): [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] self.gen.throw(typ, value, traceback) [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] raise exception.ResourceProviderSyncFailed() [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 811.518516] env[69171]: ERROR nova.compute.manager [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] [ 811.518834] env[69171]: DEBUG nova.compute.utils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 811.520217] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.052s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 811.523391] env[69171]: INFO nova.compute.claims [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 811.527972] env[69171]: DEBUG nova.compute.manager [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Build of instance 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 811.528515] env[69171]: DEBUG nova.compute.manager [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 811.528888] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Acquiring lock "refresh_cache-0c0f5347-23e3-4e69-8b9d-56eeebf39a6f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 811.528963] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Acquired lock "refresh_cache-0c0f5347-23e3-4e69-8b9d-56eeebf39a6f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 811.529121] env[69171]: DEBUG nova.network.neutron [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 811.587049] env[69171]: INFO nova.compute.manager [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] [instance: 6664e9ec-4c43-464d-9117-92967031fe55] Took 1.03 seconds to deallocate network for instance. [ 811.655334] env[69171]: DEBUG nova.compute.manager [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 812.050693] env[69171]: DEBUG nova.network.neutron [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 812.133031] env[69171]: DEBUG nova.network.neutron [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.182036] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.571035] env[69171]: DEBUG nova.scheduler.client.report [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 812.594556] env[69171]: DEBUG nova.scheduler.client.report [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 812.594556] env[69171]: DEBUG nova.compute.provider_tree [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 812.625126] env[69171]: DEBUG nova.scheduler.client.report [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 812.636608] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Releasing lock "refresh_cache-0c0f5347-23e3-4e69-8b9d-56eeebf39a6f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.638031] env[69171]: DEBUG nova.compute.manager [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 812.638031] env[69171]: DEBUG nova.compute.manager [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 812.638031] env[69171]: DEBUG nova.network.neutron [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 812.643688] env[69171]: INFO nova.scheduler.client.report [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Deleted allocations for instance 6664e9ec-4c43-464d-9117-92967031fe55 [ 812.654308] env[69171]: DEBUG nova.scheduler.client.report [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 812.661019] env[69171]: DEBUG nova.network.neutron [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 813.053752] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5336c530-757b-481f-864d-c7e0484cf8b1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.068282] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e627944c-487f-4362-aa01-6840a024e88d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.100074] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-567ef8d5-55b5-4872-8358-037828c5e1b1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.108759] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158929f8-5c7c-4d29-8bcb-a6b39a664255 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.125736] env[69171]: DEBUG nova.compute.provider_tree [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 813.157812] env[69171]: DEBUG oslo_concurrency.lockutils [None req-44dda87d-74d3-4d3f-9f7a-30fec89e7065 tempest-ServersTestBootFromVolume-875575882 tempest-ServersTestBootFromVolume-875575882-project-member] Lock "6664e9ec-4c43-464d-9117-92967031fe55" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.674s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.164869] env[69171]: DEBUG nova.network.neutron [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.661064] env[69171]: ERROR nova.scheduler.client.report [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [req-3b3b09f2-3432-4eac-948b-b59f845a39c0] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3b3b09f2-3432-4eac-948b-b59f845a39c0"}]} [ 813.661471] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.141s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.662067] env[69171]: ERROR nova.compute.manager [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 813.662067] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Traceback (most recent call last): [ 813.662067] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 813.662067] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] yield [ 813.662067] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 813.662067] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] self.set_inventory_for_provider( [ 813.662067] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 813.662067] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-3b3b09f2-3432-4eac-948b-b59f845a39c0"}]} [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] During handling of the above exception, another exception occurred: [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Traceback (most recent call last): [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] with self.rt.instance_claim(context, instance, node, allocs, [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 813.662414] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] return f(*args, **kwargs) [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] self._update(elevated, cn) [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] self._update_to_placement(context, compute_node, startup) [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] return attempt.get(self._wrap_exception) [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] raise value [ 813.662739] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] self.reportclient.update_from_provider_tree( [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] with catch_all(pd.uuid): [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] self.gen.throw(typ, value, traceback) [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] raise exception.ResourceProviderSyncFailed() [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 813.663139] env[69171]: ERROR nova.compute.manager [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] [ 813.663459] env[69171]: DEBUG nova.compute.utils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 813.666066] env[69171]: DEBUG nova.compute.manager [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 813.667487] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.171s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.669795] env[69171]: INFO nova.compute.claims [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.673706] env[69171]: DEBUG nova.compute.manager [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Build of instance aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 813.673922] env[69171]: DEBUG nova.compute.manager [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 813.674190] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Acquiring lock "refresh_cache-aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 813.674337] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Acquired lock "refresh_cache-aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 813.674493] env[69171]: DEBUG nova.network.neutron [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 813.675750] env[69171]: INFO nova.compute.manager [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] [instance: 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f] Took 1.04 seconds to deallocate network for instance. [ 814.221754] env[69171]: DEBUG nova.network.neutron [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 814.226808] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.352536] env[69171]: DEBUG nova.network.neutron [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.723313] env[69171]: DEBUG nova.scheduler.client.report [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 814.729865] env[69171]: INFO nova.scheduler.client.report [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Deleted allocations for instance 0c0f5347-23e3-4e69-8b9d-56eeebf39a6f [ 814.749896] env[69171]: DEBUG nova.scheduler.client.report [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 814.753906] env[69171]: DEBUG nova.compute.provider_tree [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 814.771288] env[69171]: DEBUG nova.scheduler.client.report [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 814.806445] env[69171]: DEBUG nova.scheduler.client.report [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 814.858747] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Releasing lock "refresh_cache-aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 814.859704] env[69171]: DEBUG nova.compute.manager [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 814.859704] env[69171]: DEBUG nova.compute.manager [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 814.859779] env[69171]: DEBUG nova.network.neutron [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 814.890027] env[69171]: DEBUG nova.network.neutron [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 815.168882] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d575779-9746-4b76-9f52-cfca0f4ae3e4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.180731] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3521485-8457-418e-939e-339b0d7f2c2f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.217657] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d9cd8f1-64d8-4fa5-b36e-314c3d568b5b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.229587] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80e4865a-b6aa-4ad0-82a9-b8f90b99a244 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.243705] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1047acdf-e0ce-4da3-acb3-b873a32ca708 tempest-ServerTagsTestJSON-1357548978 tempest-ServerTagsTestJSON-1357548978-project-member] Lock "0c0f5347-23e3-4e69-8b9d-56eeebf39a6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.383s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.244170] env[69171]: DEBUG nova.compute.provider_tree [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 815.397363] env[69171]: DEBUG nova.network.neutron [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.420343] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Acquiring lock "d950fb6b-7977-40c1-b9e1-a89bc2e9fec6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.420577] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Lock "d950fb6b-7977-40c1-b9e1-a89bc2e9fec6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.752166] env[69171]: DEBUG nova.compute.manager [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 815.775201] env[69171]: ERROR nova.scheduler.client.report [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [req-8b112a98-a1cc-4bab-b74f-1ede1e4cafdb] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8b112a98-a1cc-4bab-b74f-1ede1e4cafdb"}]} [ 815.775201] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.105s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.775320] env[69171]: ERROR nova.compute.manager [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 815.775320] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Traceback (most recent call last): [ 815.775320] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 815.775320] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] yield [ 815.775320] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 815.775320] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] self.set_inventory_for_provider( [ 815.775320] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 815.775320] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-8b112a98-a1cc-4bab-b74f-1ede1e4cafdb"}]} [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] During handling of the above exception, another exception occurred: [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Traceback (most recent call last): [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] with self.rt.instance_claim(context, instance, node, allocs, [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 815.775522] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] return f(*args, **kwargs) [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] self._update(elevated, cn) [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] self._update_to_placement(context, compute_node, startup) [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] return attempt.get(self._wrap_exception) [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] six.reraise(self.value[0], self.value[1], self.value[2]) [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] raise value [ 815.775759] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] self.reportclient.update_from_provider_tree( [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] with catch_all(pd.uuid): [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] self.gen.throw(typ, value, traceback) [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] raise exception.ResourceProviderSyncFailed() [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 815.776081] env[69171]: ERROR nova.compute.manager [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] [ 815.776375] env[69171]: DEBUG nova.compute.utils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 815.778455] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.222s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.779337] env[69171]: INFO nova.compute.claims [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 815.784136] env[69171]: DEBUG nova.compute.manager [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Build of instance 54d5f178-7fc5-48ce-be89-c783f419e581 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 815.784136] env[69171]: DEBUG nova.compute.manager [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 815.784136] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Acquiring lock "refresh_cache-54d5f178-7fc5-48ce-be89-c783f419e581" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 815.784136] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Acquired lock "refresh_cache-54d5f178-7fc5-48ce-be89-c783f419e581" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 815.784465] env[69171]: DEBUG nova.network.neutron [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 815.900624] env[69171]: INFO nova.compute.manager [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] [instance: aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3] Took 1.04 seconds to deallocate network for instance. [ 816.281319] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.316378] env[69171]: DEBUG nova.network.neutron [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 816.419864] env[69171]: DEBUG nova.network.neutron [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.710861] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Acquiring lock "13d5116e-46ab-40d4-a507-3d34fedd7d4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.711243] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Lock "13d5116e-46ab-40d4-a507-3d34fedd7d4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.818348] env[69171]: DEBUG nova.scheduler.client.report [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 816.831175] env[69171]: DEBUG nova.scheduler.client.report [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 816.831429] env[69171]: DEBUG nova.compute.provider_tree [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 816.843847] env[69171]: DEBUG nova.scheduler.client.report [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 816.862423] env[69171]: DEBUG nova.scheduler.client.report [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 816.918853] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Releasing lock "refresh_cache-54d5f178-7fc5-48ce-be89-c783f419e581" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 816.919550] env[69171]: DEBUG nova.compute.manager [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 816.919550] env[69171]: DEBUG nova.compute.manager [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 816.919550] env[69171]: DEBUG nova.network.neutron [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 816.934896] env[69171]: INFO nova.scheduler.client.report [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Deleted allocations for instance aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3 [ 816.941750] env[69171]: DEBUG nova.network.neutron [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 817.232307] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-118a0469-fb8a-443a-8eb6-6246c6ebb9c1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.243023] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a60183-ce1b-4984-9363-adfcd9719271 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.279597] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-435f3825-e799-474f-b0ee-e594ce53babb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.286270] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0efb5f15-7474-49dd-b992-9b6960db7a40 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.302212] env[69171]: DEBUG nova.compute.provider_tree [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 817.446373] env[69171]: DEBUG nova.network.neutron [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.447870] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3a61a5e1-63f8-4c88-abdc-2f2abe2781b4 tempest-ServersTestJSON-898140240 tempest-ServersTestJSON-898140240-project-member] Lock "aee814d6-dfc7-4ea9-84ec-585c6d5ea0e3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.006s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.831339] env[69171]: ERROR nova.scheduler.client.report [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [req-0f3b7e76-37b4-4d7c-a860-d51d2f171551] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0f3b7e76-37b4-4d7c-a860-d51d2f171551"}]} [ 817.831339] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.052s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.831639] env[69171]: ERROR nova.compute.manager [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 817.831639] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Traceback (most recent call last): [ 817.831639] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 817.831639] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] yield [ 817.831639] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 817.831639] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] self.set_inventory_for_provider( [ 817.831639] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 817.831639] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0f3b7e76-37b4-4d7c-a860-d51d2f171551"}]} [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] During handling of the above exception, another exception occurred: [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Traceback (most recent call last): [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] with self.rt.instance_claim(context, instance, node, allocs, [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 817.831895] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] return f(*args, **kwargs) [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] self._update(elevated, cn) [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] self._update_to_placement(context, compute_node, startup) [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] return attempt.get(self._wrap_exception) [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] raise value [ 817.832186] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] self.reportclient.update_from_provider_tree( [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] with catch_all(pd.uuid): [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] self.gen.throw(typ, value, traceback) [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] raise exception.ResourceProviderSyncFailed() [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 817.832610] env[69171]: ERROR nova.compute.manager [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] [ 817.832911] env[69171]: DEBUG nova.compute.utils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 817.832911] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.920s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.834607] env[69171]: INFO nova.compute.claims [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.836697] env[69171]: DEBUG nova.compute.manager [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Build of instance 478e1a26-101c-48a4-bcc2-a103f59962f6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 817.837134] env[69171]: DEBUG nova.compute.manager [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 817.837421] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Acquiring lock "refresh_cache-478e1a26-101c-48a4-bcc2-a103f59962f6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.837508] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Acquired lock "refresh_cache-478e1a26-101c-48a4-bcc2-a103f59962f6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.838505] env[69171]: DEBUG nova.network.neutron [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 817.951477] env[69171]: INFO nova.compute.manager [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] [instance: 54d5f178-7fc5-48ce-be89-c783f419e581] Took 1.03 seconds to deallocate network for instance. [ 817.955012] env[69171]: DEBUG nova.compute.manager [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 818.369163] env[69171]: DEBUG nova.network.neutron [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 818.490277] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 818.529292] env[69171]: DEBUG nova.network.neutron [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.870412] env[69171]: DEBUG nova.scheduler.client.report [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 818.891763] env[69171]: DEBUG nova.scheduler.client.report [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 818.891763] env[69171]: DEBUG nova.compute.provider_tree [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 818.907919] env[69171]: DEBUG nova.scheduler.client.report [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 818.929466] env[69171]: DEBUG nova.scheduler.client.report [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 818.990225] env[69171]: INFO nova.scheduler.client.report [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Deleted allocations for instance 54d5f178-7fc5-48ce-be89-c783f419e581 [ 819.030721] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Releasing lock "refresh_cache-478e1a26-101c-48a4-bcc2-a103f59962f6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.030924] env[69171]: DEBUG nova.compute.manager [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 819.031124] env[69171]: DEBUG nova.compute.manager [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 819.031319] env[69171]: DEBUG nova.network.neutron [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 819.055446] env[69171]: DEBUG nova.network.neutron [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 819.321256] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3ad6f9e-3d67-4efb-bde9-b92472d99720 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.331079] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5f2fc57-ec73-456e-8eb3-aeb2d2dd54f7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.365072] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824a4d8f-df37-46ee-b4ca-ed7a9c0c40fc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.374136] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a36bb9-8dca-4cfd-9db0-33c6c86e3d2d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.389928] env[69171]: DEBUG nova.compute.provider_tree [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 819.496507] env[69171]: DEBUG oslo_concurrency.lockutils [None req-07a2deed-d8da-4be5-a632-2d26c47af83b tempest-ServerActionsV293TestJSON-1367409092 tempest-ServerActionsV293TestJSON-1367409092-project-member] Lock "54d5f178-7fc5-48ce-be89-c783f419e581" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.125s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.557849] env[69171]: DEBUG nova.network.neutron [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.915430] env[69171]: ERROR nova.scheduler.client.report [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [req-0bed1267-7a78-442a-98a1-a4aa1d007d4f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0bed1267-7a78-442a-98a1-a4aa1d007d4f"}]} [ 819.915797] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.084s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.916403] env[69171]: ERROR nova.compute.manager [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 819.916403] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Traceback (most recent call last): [ 819.916403] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 819.916403] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] yield [ 819.916403] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 819.916403] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] self.set_inventory_for_provider( [ 819.916403] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 819.916403] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0bed1267-7a78-442a-98a1-a4aa1d007d4f"}]} [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] During handling of the above exception, another exception occurred: [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Traceback (most recent call last): [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] with self.rt.instance_claim(context, instance, node, allocs, [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 819.916714] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] return f(*args, **kwargs) [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] self._update(elevated, cn) [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] self._update_to_placement(context, compute_node, startup) [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] return attempt.get(self._wrap_exception) [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] six.reraise(self.value[0], self.value[1], self.value[2]) [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] raise value [ 819.917061] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] self.reportclient.update_from_provider_tree( [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] with catch_all(pd.uuid): [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] self.gen.throw(typ, value, traceback) [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] raise exception.ResourceProviderSyncFailed() [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 819.917441] env[69171]: ERROR nova.compute.manager [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] [ 819.917751] env[69171]: DEBUG nova.compute.utils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 819.920763] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.944s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.922545] env[69171]: INFO nova.compute.claims [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 819.925529] env[69171]: DEBUG nova.compute.manager [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Build of instance ed037ce0-072f-425a-97fa-5fb0f941268b was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 819.929272] env[69171]: DEBUG nova.compute.manager [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 819.929585] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Acquiring lock "refresh_cache-ed037ce0-072f-425a-97fa-5fb0f941268b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 819.929738] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Acquired lock "refresh_cache-ed037ce0-072f-425a-97fa-5fb0f941268b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.929907] env[69171]: DEBUG nova.network.neutron [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 820.000318] env[69171]: DEBUG nova.compute.manager [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 820.063125] env[69171]: INFO nova.compute.manager [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] [instance: 478e1a26-101c-48a4-bcc2-a103f59962f6] Took 1.03 seconds to deallocate network for instance. [ 820.203102] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquiring lock "0da1037f-5468-4926-be0a-8a91f31f8bdb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.204179] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "0da1037f-5468-4926-be0a-8a91f31f8bdb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.456272] env[69171]: DEBUG nova.network.neutron [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 820.532226] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.587318] env[69171]: DEBUG nova.network.neutron [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.885479] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquiring lock "87cec11c-66c7-4b71-946b-1e8e7d72f9aa" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.887310] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "87cec11c-66c7-4b71-946b-1e8e7d72f9aa" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.962155] env[69171]: DEBUG nova.scheduler.client.report [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 820.986730] env[69171]: DEBUG nova.scheduler.client.report [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 820.986834] env[69171]: DEBUG nova.compute.provider_tree [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.006408] env[69171]: DEBUG nova.scheduler.client.report [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 821.032578] env[69171]: DEBUG nova.scheduler.client.report [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 821.094051] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Releasing lock "refresh_cache-ed037ce0-072f-425a-97fa-5fb0f941268b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.094253] env[69171]: DEBUG nova.compute.manager [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 821.095030] env[69171]: DEBUG nova.compute.manager [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 821.095030] env[69171]: DEBUG nova.network.neutron [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 821.114221] env[69171]: INFO nova.scheduler.client.report [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Deleted allocations for instance 478e1a26-101c-48a4-bcc2-a103f59962f6 [ 821.124167] env[69171]: DEBUG nova.network.neutron [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 821.543271] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b81a2e-0a9f-4e41-8e0e-adca60cc6795 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.553282] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9fa3c2-17b3-40df-8000-e711319ca0d6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.587192] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d9fc2eb-61f9-44d5-8c3c-c559d1919c20 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.592840] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Acquiring lock "6071a99a-ff47-4161-8a1e-65385318891d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.593328] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Lock "6071a99a-ff47-4161-8a1e-65385318891d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.599146] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f6ff23-4f71-401d-b002-9e7e7f4d2194 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.617075] env[69171]: DEBUG nova.compute.provider_tree [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 821.627154] env[69171]: DEBUG nova.network.neutron [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.628440] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1c2200c4-d540-449f-9004-0cd649a7c2b9 tempest-ServerMetadataNegativeTestJSON-999181609 tempest-ServerMetadataNegativeTestJSON-999181609-project-member] Lock "478e1a26-101c-48a4-bcc2-a103f59962f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.059s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.129553] env[69171]: INFO nova.compute.manager [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] [instance: ed037ce0-072f-425a-97fa-5fb0f941268b] Took 1.03 seconds to deallocate network for instance. [ 822.132927] env[69171]: DEBUG nova.compute.manager [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 822.152635] env[69171]: ERROR nova.scheduler.client.report [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [req-210ce461-4496-4ee6-8d92-f2c4f980dd00] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-210ce461-4496-4ee6-8d92-f2c4f980dd00"}]} [ 822.153031] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.232s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 822.153665] env[69171]: ERROR nova.compute.manager [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 822.153665] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Traceback (most recent call last): [ 822.153665] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 822.153665] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] yield [ 822.153665] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 822.153665] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] self.set_inventory_for_provider( [ 822.153665] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 822.153665] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-210ce461-4496-4ee6-8d92-f2c4f980dd00"}]} [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] During handling of the above exception, another exception occurred: [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Traceback (most recent call last): [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] with self.rt.instance_claim(context, instance, node, allocs, [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 822.153893] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] return f(*args, **kwargs) [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] self._update(elevated, cn) [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] self._update_to_placement(context, compute_node, startup) [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] return attempt.get(self._wrap_exception) [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] six.reraise(self.value[0], self.value[1], self.value[2]) [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] raise value [ 822.154187] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] self.reportclient.update_from_provider_tree( [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] with catch_all(pd.uuid): [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] self.gen.throw(typ, value, traceback) [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] raise exception.ResourceProviderSyncFailed() [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 822.154784] env[69171]: ERROR nova.compute.manager [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] [ 822.155099] env[69171]: DEBUG nova.compute.utils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 822.157902] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.065s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.160014] env[69171]: INFO nova.compute.claims [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 822.162624] env[69171]: DEBUG nova.compute.manager [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Build of instance 776e2ab5-db02-46c8-a971-08b3c15e903f was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 822.163225] env[69171]: DEBUG nova.compute.manager [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 822.163325] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Acquiring lock "refresh_cache-776e2ab5-db02-46c8-a971-08b3c15e903f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.163480] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Acquired lock "refresh_cache-776e2ab5-db02-46c8-a971-08b3c15e903f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 822.163637] env[69171]: DEBUG nova.network.neutron [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 822.663129] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.690203] env[69171]: DEBUG nova.network.neutron [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 822.829060] env[69171]: DEBUG nova.network.neutron [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.176659] env[69171]: INFO nova.scheduler.client.report [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Deleted allocations for instance ed037ce0-072f-425a-97fa-5fb0f941268b [ 823.203412] env[69171]: DEBUG nova.scheduler.client.report [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 823.227513] env[69171]: DEBUG nova.scheduler.client.report [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 823.228044] env[69171]: DEBUG nova.compute.provider_tree [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 823.242565] env[69171]: DEBUG nova.scheduler.client.report [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 823.272258] env[69171]: DEBUG nova.scheduler.client.report [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 823.332764] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Releasing lock "refresh_cache-776e2ab5-db02-46c8-a971-08b3c15e903f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.332764] env[69171]: DEBUG nova.compute.manager [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 823.332764] env[69171]: DEBUG nova.compute.manager [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 823.332954] env[69171]: DEBUG nova.network.neutron [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 823.372643] env[69171]: DEBUG nova.network.neutron [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 823.689477] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8128d36c-ccb4-4008-bee8-32b92d364c9e tempest-ServerActionsTestJSON-2043156247 tempest-ServerActionsTestJSON-2043156247-project-member] Lock "ed037ce0-072f-425a-97fa-5fb0f941268b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.798s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.707220] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51039533-5b19-4583-aff3-1eefa3b6b48f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.719354] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07f5e965-3aed-43c2-840f-df2e23012ca0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.769054] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4f4ca54-f520-4aad-83f9-a4610f2b152c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.775438] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb4740f1-e245-46db-85b2-653a1a4a7c9f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.802708] env[69171]: DEBUG nova.compute.provider_tree [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 823.880824] env[69171]: DEBUG nova.network.neutron [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.192537] env[69171]: DEBUG nova.compute.manager [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 824.339517] env[69171]: ERROR nova.scheduler.client.report [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [req-9d3651a9-a5e9-451f-a31b-f43a774b46bd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9d3651a9-a5e9-451f-a31b-f43a774b46bd"}]} [ 824.339517] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.181s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.339871] env[69171]: ERROR nova.compute.manager [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 824.339871] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Traceback (most recent call last): [ 824.339871] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 824.339871] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] yield [ 824.339871] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 824.339871] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] self.set_inventory_for_provider( [ 824.339871] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 824.339871] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-9d3651a9-a5e9-451f-a31b-f43a774b46bd"}]} [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] During handling of the above exception, another exception occurred: [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Traceback (most recent call last): [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] with self.rt.instance_claim(context, instance, node, allocs, [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 824.340132] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] return f(*args, **kwargs) [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] self._update(elevated, cn) [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] self._update_to_placement(context, compute_node, startup) [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] return attempt.get(self._wrap_exception) [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] six.reraise(self.value[0], self.value[1], self.value[2]) [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] raise value [ 824.340515] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] self.reportclient.update_from_provider_tree( [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] with catch_all(pd.uuid): [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] self.gen.throw(typ, value, traceback) [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] raise exception.ResourceProviderSyncFailed() [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 824.342067] env[69171]: ERROR nova.compute.manager [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] [ 824.342516] env[69171]: DEBUG nova.compute.utils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 824.342977] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.169s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 824.345052] env[69171]: INFO nova.compute.claims [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.348767] env[69171]: DEBUG nova.compute.manager [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Build of instance f13b2145-5f8b-4b0a-ac07-c50018082fe2 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 824.349506] env[69171]: DEBUG nova.compute.manager [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 824.349506] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "refresh_cache-f13b2145-5f8b-4b0a-ac07-c50018082fe2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 824.349506] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquired lock "refresh_cache-f13b2145-5f8b-4b0a-ac07-c50018082fe2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 824.349506] env[69171]: DEBUG nova.network.neutron [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 824.383975] env[69171]: INFO nova.compute.manager [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] [instance: 776e2ab5-db02-46c8-a971-08b3c15e903f] Took 1.05 seconds to deallocate network for instance. [ 824.718742] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.883138] env[69171]: DEBUG nova.network.neutron [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.023435] env[69171]: DEBUG nova.network.neutron [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.389584] env[69171]: DEBUG nova.scheduler.client.report [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 825.418172] env[69171]: DEBUG nova.scheduler.client.report [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 825.418741] env[69171]: DEBUG nova.compute.provider_tree [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.432700] env[69171]: INFO nova.scheduler.client.report [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Deleted allocations for instance 776e2ab5-db02-46c8-a971-08b3c15e903f [ 825.439273] env[69171]: DEBUG nova.scheduler.client.report [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 825.465926] env[69171]: DEBUG nova.scheduler.client.report [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 825.526316] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Releasing lock "refresh_cache-f13b2145-5f8b-4b0a-ac07-c50018082fe2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.526594] env[69171]: DEBUG nova.compute.manager [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 825.526857] env[69171]: DEBUG nova.compute.manager [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 825.530066] env[69171]: DEBUG nova.network.neutron [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 825.557524] env[69171]: DEBUG nova.network.neutron [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 825.857954] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c466fb1d-10e5-4f76-af94-eaeee14157da {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.866758] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-386bf49b-18d5-4076-ba55-6400ce0b645b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.899778] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-583330b3-f5b8-40cb-a107-e7f3a75579e2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.908636] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f86715ac-c755-4e92-b016-838481fe9e4b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.925704] env[69171]: DEBUG nova.compute.provider_tree [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 825.943612] env[69171]: DEBUG oslo_concurrency.lockutils [None req-099884b4-8a5a-4532-bfe8-e1e848d77641 tempest-ServersTestFqdnHostnames-1379955857 tempest-ServersTestFqdnHostnames-1379955857-project-member] Lock "776e2ab5-db02-46c8-a971-08b3c15e903f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.354s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.061241] env[69171]: DEBUG nova.network.neutron [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.410515] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Acquiring lock "6a0befb0-adf1-4907-845b-7e29f28f4f6a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.410787] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Lock "6a0befb0-adf1-4907-845b-7e29f28f4f6a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.448998] env[69171]: DEBUG nova.compute.manager [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 826.456639] env[69171]: ERROR nova.scheduler.client.report [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [req-f0e5487f-e62f-4a0a-9ed5-015c686f309d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f0e5487f-e62f-4a0a-9ed5-015c686f309d"}]} [ 826.456639] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.112s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.456935] env[69171]: ERROR nova.compute.manager [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 826.456935] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Traceback (most recent call last): [ 826.456935] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 826.456935] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] yield [ 826.456935] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 826.456935] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] self.set_inventory_for_provider( [ 826.456935] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 826.456935] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f0e5487f-e62f-4a0a-9ed5-015c686f309d"}]} [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] During handling of the above exception, another exception occurred: [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Traceback (most recent call last): [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] with self.rt.instance_claim(context, instance, node, allocs, [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 826.457159] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] return f(*args, **kwargs) [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] self._update(elevated, cn) [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] self._update_to_placement(context, compute_node, startup) [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] return attempt.get(self._wrap_exception) [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] six.reraise(self.value[0], self.value[1], self.value[2]) [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] raise value [ 826.457426] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] self.reportclient.update_from_provider_tree( [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] with catch_all(pd.uuid): [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] self.gen.throw(typ, value, traceback) [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] raise exception.ResourceProviderSyncFailed() [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 826.457764] env[69171]: ERROR nova.compute.manager [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] [ 826.458062] env[69171]: DEBUG nova.compute.utils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 826.458062] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.276s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.458922] env[69171]: INFO nova.compute.claims [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.464049] env[69171]: DEBUG nova.compute.manager [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Build of instance f0af7537-ef12-4a21-8c4a-8afc46be9e0b was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 826.464049] env[69171]: DEBUG nova.compute.manager [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 826.464049] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "refresh_cache-f0af7537-ef12-4a21-8c4a-8afc46be9e0b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.464049] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquired lock "refresh_cache-f0af7537-ef12-4a21-8c4a-8afc46be9e0b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.464049] env[69171]: DEBUG nova.network.neutron [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 826.564383] env[69171]: INFO nova.compute.manager [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: f13b2145-5f8b-4b0a-ac07-c50018082fe2] Took 1.04 seconds to deallocate network for instance. [ 826.980585] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.997982] env[69171]: DEBUG nova.network.neutron [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 827.227295] env[69171]: DEBUG nova.network.neutron [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.491033] env[69171]: DEBUG nova.scheduler.client.report [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 827.511095] env[69171]: DEBUG nova.scheduler.client.report [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 827.511095] env[69171]: DEBUG nova.compute.provider_tree [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 827.526670] env[69171]: DEBUG nova.scheduler.client.report [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 827.556129] env[69171]: DEBUG nova.scheduler.client.report [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 827.596376] env[69171]: INFO nova.scheduler.client.report [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Deleted allocations for instance f13b2145-5f8b-4b0a-ac07-c50018082fe2 [ 827.731176] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Releasing lock "refresh_cache-f0af7537-ef12-4a21-8c4a-8afc46be9e0b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.731176] env[69171]: DEBUG nova.compute.manager [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 827.732320] env[69171]: DEBUG nova.compute.manager [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 827.732320] env[69171]: DEBUG nova.network.neutron [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 827.763746] env[69171]: DEBUG nova.network.neutron [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 828.035279] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb59213-49b5-4dae-8f64-eeb7c54dff45 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.050502] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac5d8f4c-6e54-459e-bea6-2c0685a5489a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.084165] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8310b443-9c6b-4288-bff8-f32a33c82014 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.092210] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7422474-7d3e-4eac-a39e-e262fade6e9f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.107011] env[69171]: DEBUG nova.compute.provider_tree [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 828.110146] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e4dbbae6-b184-4f54-bca6-82d8d26c8c76 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "f13b2145-5f8b-4b0a-ac07-c50018082fe2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.381s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.267635] env[69171]: DEBUG nova.network.neutron [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.613159] env[69171]: DEBUG nova.compute.manager [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 828.636472] env[69171]: ERROR nova.scheduler.client.report [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [req-22895c5a-1178-444c-aa0b-49f271d3b97a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-22895c5a-1178-444c-aa0b-49f271d3b97a"}]} [ 828.636472] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.179s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.638524] env[69171]: ERROR nova.compute.manager [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 828.638524] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Traceback (most recent call last): [ 828.638524] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 828.638524] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] yield [ 828.638524] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 828.638524] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] self.set_inventory_for_provider( [ 828.638524] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 828.638524] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-22895c5a-1178-444c-aa0b-49f271d3b97a"}]} [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] During handling of the above exception, another exception occurred: [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Traceback (most recent call last): [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] with self.rt.instance_claim(context, instance, node, allocs, [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 828.640085] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] return f(*args, **kwargs) [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] self._update(elevated, cn) [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] self._update_to_placement(context, compute_node, startup) [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] return attempt.get(self._wrap_exception) [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] raise value [ 828.640393] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] self.reportclient.update_from_provider_tree( [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] with catch_all(pd.uuid): [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] self.gen.throw(typ, value, traceback) [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] raise exception.ResourceProviderSyncFailed() [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 828.640756] env[69171]: ERROR nova.compute.manager [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] [ 828.641168] env[69171]: DEBUG nova.compute.utils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 828.641168] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.413s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.641885] env[69171]: INFO nova.compute.claims [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.644859] env[69171]: DEBUG nova.compute.manager [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Build of instance aed06808-707f-4d80-ad42-bdd2e8f90cd6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 828.646030] env[69171]: DEBUG nova.compute.manager [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 828.646030] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquiring lock "refresh_cache-aed06808-707f-4d80-ad42-bdd2e8f90cd6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.646270] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Acquired lock "refresh_cache-aed06808-707f-4d80-ad42-bdd2e8f90cd6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.646547] env[69171]: DEBUG nova.network.neutron [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 828.772814] env[69171]: INFO nova.compute.manager [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: f0af7537-ef12-4a21-8c4a-8afc46be9e0b] Took 1.04 seconds to deallocate network for instance. [ 829.148037] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.184234] env[69171]: DEBUG nova.network.neutron [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 829.370970] env[69171]: DEBUG nova.network.neutron [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.668397] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Acquiring lock "1875b4bd-9024-40c0-9c43-244224904ce9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.668653] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Lock "1875b4bd-9024-40c0-9c43-244224904ce9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.682952] env[69171]: DEBUG nova.scheduler.client.report [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 829.707146] env[69171]: DEBUG nova.scheduler.client.report [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 829.707146] env[69171]: DEBUG nova.compute.provider_tree [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 829.725126] env[69171]: DEBUG nova.scheduler.client.report [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 829.751751] env[69171]: DEBUG nova.scheduler.client.report [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 829.810959] env[69171]: INFO nova.scheduler.client.report [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Deleted allocations for instance f0af7537-ef12-4a21-8c4a-8afc46be9e0b [ 829.873847] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Releasing lock "refresh_cache-aed06808-707f-4d80-ad42-bdd2e8f90cd6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.874092] env[69171]: DEBUG nova.compute.manager [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 829.874283] env[69171]: DEBUG nova.compute.manager [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 829.874454] env[69171]: DEBUG nova.network.neutron [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 829.894798] env[69171]: DEBUG nova.network.neutron [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 830.146022] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77989b72-a40a-49ef-8b2a-159ee6c541c6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.152097] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23248442-f1bd-4572-b005-69285118f2eb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.186200] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37e73753-b066-47bb-8532-23c827cf3e5d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.195922] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d911e5c9-4bb0-4d0c-b5df-d013244d7d1d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.220300] env[69171]: DEBUG nova.compute.provider_tree [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 830.320655] env[69171]: DEBUG oslo_concurrency.lockutils [None req-699b2be1-c92e-4403-a6a1-4391774faff9 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "f0af7537-ef12-4a21-8c4a-8afc46be9e0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.748s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.397824] env[69171]: DEBUG nova.network.neutron [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.741130] env[69171]: ERROR nova.scheduler.client.report [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [req-5bda0feb-c883-4d88-99e7-72c8b40e1cec] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5bda0feb-c883-4d88-99e7-72c8b40e1cec"}]} [ 830.741574] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.102s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.742131] env[69171]: ERROR nova.compute.manager [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 830.742131] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Traceback (most recent call last): [ 830.742131] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 830.742131] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] yield [ 830.742131] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 830.742131] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] self.set_inventory_for_provider( [ 830.742131] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 830.742131] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5bda0feb-c883-4d88-99e7-72c8b40e1cec"}]} [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] During handling of the above exception, another exception occurred: [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Traceback (most recent call last): [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] with self.rt.instance_claim(context, instance, node, allocs, [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 830.742366] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] return f(*args, **kwargs) [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] self._update(elevated, cn) [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] self._update_to_placement(context, compute_node, startup) [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] return attempt.get(self._wrap_exception) [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] six.reraise(self.value[0], self.value[1], self.value[2]) [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] raise value [ 830.742681] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] self.reportclient.update_from_provider_tree( [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] with catch_all(pd.uuid): [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] self.gen.throw(typ, value, traceback) [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] raise exception.ResourceProviderSyncFailed() [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 830.743052] env[69171]: ERROR nova.compute.manager [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] [ 830.743337] env[69171]: DEBUG nova.compute.utils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 830.744047] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.463s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.745504] env[69171]: INFO nova.compute.claims [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.748172] env[69171]: DEBUG nova.compute.manager [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Build of instance 211d6fa6-6d34-45e8-bfc3-b5feccb642de was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 830.748579] env[69171]: DEBUG nova.compute.manager [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 830.748804] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "refresh_cache-211d6fa6-6d34-45e8-bfc3-b5feccb642de" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 830.748951] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquired lock "refresh_cache-211d6fa6-6d34-45e8-bfc3-b5feccb642de" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 830.749239] env[69171]: DEBUG nova.network.neutron [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 830.824854] env[69171]: DEBUG nova.compute.manager [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 830.900729] env[69171]: INFO nova.compute.manager [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] [instance: aed06808-707f-4d80-ad42-bdd2e8f90cd6] Took 1.03 seconds to deallocate network for instance. [ 831.270685] env[69171]: DEBUG nova.network.neutron [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.314568] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "de9e4934-543e-433b-9094-ef2ce264a2f6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.314809] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "de9e4934-543e-433b-9094-ef2ce264a2f6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.319015] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.319266] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.347464] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.359096] env[69171]: DEBUG nova.network.neutron [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 831.774779] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 831.788279] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 831.788547] env[69171]: DEBUG nova.compute.provider_tree [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 831.800490] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 831.819035] env[69171]: DEBUG nova.scheduler.client.report [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 831.861731] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Releasing lock "refresh_cache-211d6fa6-6d34-45e8-bfc3-b5feccb642de" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 831.861957] env[69171]: DEBUG nova.compute.manager [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 831.862265] env[69171]: DEBUG nova.compute.manager [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 831.862501] env[69171]: DEBUG nova.network.neutron [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 831.879074] env[69171]: DEBUG nova.network.neutron [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 831.928118] env[69171]: INFO nova.scheduler.client.report [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Deleted allocations for instance aed06808-707f-4d80-ad42-bdd2e8f90cd6 [ 832.144307] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3935fb4-7054-4edf-be09-c3435f76f31e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.219314] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b92c48d-ff9b-4679-bd36-121ba018634b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.288396] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7e68192-c55c-4455-a4e6-12e2f2448dac {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.307066] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cef903a-7ba0-48d0-aac9-c91f1f10d35b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.341362] env[69171]: DEBUG nova.compute.provider_tree [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.384328] env[69171]: DEBUG nova.network.neutron [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.438237] env[69171]: DEBUG oslo_concurrency.lockutils [None req-1ac6412a-ea20-421d-9045-e0177b912dc4 tempest-SecurityGroupsTestJSON-1011558079 tempest-SecurityGroupsTestJSON-1011558079-project-member] Lock "aed06808-707f-4d80-ad42-bdd2e8f90cd6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.239s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.871947] env[69171]: ERROR nova.scheduler.client.report [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [req-d66abd22-c0f9-48f7-a5c7-ae04d80daa6e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d66abd22-c0f9-48f7-a5c7-ae04d80daa6e"}]} [ 832.872356] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.128s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.873014] env[69171]: ERROR nova.compute.manager [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 832.873014] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Traceback (most recent call last): [ 832.873014] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 832.873014] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] yield [ 832.873014] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 832.873014] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] self.set_inventory_for_provider( [ 832.873014] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 832.873014] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d66abd22-c0f9-48f7-a5c7-ae04d80daa6e"}]} [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] During handling of the above exception, another exception occurred: [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Traceback (most recent call last): [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] with self.rt.instance_claim(context, instance, node, allocs, [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 832.873298] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] return f(*args, **kwargs) [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] self._update(elevated, cn) [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] self._update_to_placement(context, compute_node, startup) [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] return attempt.get(self._wrap_exception) [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] six.reraise(self.value[0], self.value[1], self.value[2]) [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] raise value [ 832.873675] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] self.reportclient.update_from_provider_tree( [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] with catch_all(pd.uuid): [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] self.gen.throw(typ, value, traceback) [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] raise exception.ResourceProviderSyncFailed() [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 832.874114] env[69171]: ERROR nova.compute.manager [instance: c0976d99-04aa-4685-8408-ff5e69de129f] [ 832.874482] env[69171]: DEBUG nova.compute.utils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 832.875010] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.385s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.876466] env[69171]: INFO nova.compute.claims [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 832.879056] env[69171]: DEBUG nova.compute.manager [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Build of instance c0976d99-04aa-4685-8408-ff5e69de129f was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 832.879527] env[69171]: DEBUG nova.compute.manager [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 832.879688] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquiring lock "refresh_cache-c0976d99-04aa-4685-8408-ff5e69de129f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.879840] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquired lock "refresh_cache-c0976d99-04aa-4685-8408-ff5e69de129f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.879994] env[69171]: DEBUG nova.network.neutron [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 832.886447] env[69171]: INFO nova.compute.manager [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 211d6fa6-6d34-45e8-bfc3-b5feccb642de] Took 1.02 seconds to deallocate network for instance. [ 832.940767] env[69171]: DEBUG nova.compute.manager [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 833.403874] env[69171]: DEBUG nova.network.neutron [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 833.465485] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.466419] env[69171]: DEBUG nova.network.neutron [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.910348] env[69171]: DEBUG nova.scheduler.client.report [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 833.918089] env[69171]: INFO nova.scheduler.client.report [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Deleted allocations for instance 211d6fa6-6d34-45e8-bfc3-b5feccb642de [ 833.925305] env[69171]: DEBUG nova.scheduler.client.report [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 833.925578] env[69171]: DEBUG nova.compute.provider_tree [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 833.952014] env[69171]: DEBUG nova.scheduler.client.report [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 833.968563] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Releasing lock "refresh_cache-c0976d99-04aa-4685-8408-ff5e69de129f" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.968807] env[69171]: DEBUG nova.compute.manager [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 833.968974] env[69171]: DEBUG nova.compute.manager [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 833.969163] env[69171]: DEBUG nova.network.neutron [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 833.976032] env[69171]: DEBUG nova.scheduler.client.report [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 834.002938] env[69171]: DEBUG nova.network.neutron [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 834.283066] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c3210d4-74d1-4ce6-a4bd-a3a091f57c9f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.291290] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-093b2061-04aa-4ac7-a509-113fc8762e71 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.323115] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd572b9-d31a-4e38-a9e4-776af0cf4e6e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.331333] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-853fdf18-d3f8-4e68-b1d0-e417156b1de7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.345804] env[69171]: DEBUG nova.compute.provider_tree [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 834.432560] env[69171]: DEBUG oslo_concurrency.lockutils [None req-dba42eab-88ea-4828-aeb7-52da045350b4 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "211d6fa6-6d34-45e8-bfc3-b5feccb642de" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.152s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.508512] env[69171]: DEBUG nova.network.neutron [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.874396] env[69171]: ERROR nova.scheduler.client.report [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [req-1b85f7af-05af-465e-b029-257e0b3fce14] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1b85f7af-05af-465e-b029-257e0b3fce14"}]} [ 834.874729] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.875349] env[69171]: ERROR nova.compute.manager [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 834.875349] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Traceback (most recent call last): [ 834.875349] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 834.875349] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] yield [ 834.875349] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 834.875349] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] self.set_inventory_for_provider( [ 834.875349] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 834.875349] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1b85f7af-05af-465e-b029-257e0b3fce14"}]} [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] During handling of the above exception, another exception occurred: [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Traceback (most recent call last): [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] with self.rt.instance_claim(context, instance, node, allocs, [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 834.875633] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] return f(*args, **kwargs) [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] self._update(elevated, cn) [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] self._update_to_placement(context, compute_node, startup) [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] return attempt.get(self._wrap_exception) [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] six.reraise(self.value[0], self.value[1], self.value[2]) [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] raise value [ 834.875972] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] self.reportclient.update_from_provider_tree( [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] with catch_all(pd.uuid): [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] self.gen.throw(typ, value, traceback) [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] raise exception.ResourceProviderSyncFailed() [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 834.876425] env[69171]: ERROR nova.compute.manager [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] [ 834.876827] env[69171]: DEBUG nova.compute.utils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 834.877351] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.345s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.878789] env[69171]: INFO nova.compute.claims [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 834.881576] env[69171]: DEBUG nova.compute.manager [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Build of instance b189c5ed-d475-44ec-b26e-e7c7014f0a81 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 834.882012] env[69171]: DEBUG nova.compute.manager [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 834.882257] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "refresh_cache-b189c5ed-d475-44ec-b26e-e7c7014f0a81" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.882454] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquired lock "refresh_cache-b189c5ed-d475-44ec-b26e-e7c7014f0a81" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.882618] env[69171]: DEBUG nova.network.neutron [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 834.935589] env[69171]: DEBUG nova.compute.manager [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 835.013167] env[69171]: INFO nova.compute.manager [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c0976d99-04aa-4685-8408-ff5e69de129f] Took 1.04 seconds to deallocate network for instance. [ 835.410957] env[69171]: DEBUG nova.network.neutron [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 835.467310] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.505731] env[69171]: DEBUG nova.network.neutron [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.909849] env[69171]: DEBUG nova.scheduler.client.report [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 835.926590] env[69171]: DEBUG nova.scheduler.client.report [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 835.926590] env[69171]: DEBUG nova.compute.provider_tree [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 835.939368] env[69171]: DEBUG nova.scheduler.client.report [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 835.956066] env[69171]: DEBUG nova.scheduler.client.report [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 836.008748] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Releasing lock "refresh_cache-b189c5ed-d475-44ec-b26e-e7c7014f0a81" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.008991] env[69171]: DEBUG nova.compute.manager [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 836.009191] env[69171]: DEBUG nova.compute.manager [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 836.009381] env[69171]: DEBUG nova.network.neutron [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 836.031079] env[69171]: DEBUG nova.network.neutron [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 836.038703] env[69171]: INFO nova.scheduler.client.report [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Deleted allocations for instance c0976d99-04aa-4685-8408-ff5e69de129f [ 836.261311] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71ba6239-0862-4937-ad2a-376211ec910a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.271785] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9872d308-6e18-4faf-b9a2-738457dde402 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.310728] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a210bf6e-6933-4993-aff3-98afbf2efb18 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.321025] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7168cdd-d2a1-42fb-b01c-94e84241ba06 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.336322] env[69171]: DEBUG nova.compute.provider_tree [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 836.526486] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "44295293-0089-49c8-bdda-97ba3331c7d0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 836.526720] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "44295293-0089-49c8-bdda-97ba3331c7d0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.533500] env[69171]: DEBUG nova.network.neutron [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.550258] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a6b3b613-105f-4fcd-9c95-bb0c54aa0161 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "c0976d99-04aa-4685-8408-ff5e69de129f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.104s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.858046] env[69171]: ERROR nova.scheduler.client.report [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [req-b002ed5b-ba60-4692-a7bf-4e5cf941810e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b002ed5b-ba60-4692-a7bf-4e5cf941810e"}]} [ 836.858482] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.981s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.859126] env[69171]: ERROR nova.compute.manager [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 836.859126] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Traceback (most recent call last): [ 836.859126] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 836.859126] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] yield [ 836.859126] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 836.859126] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] self.set_inventory_for_provider( [ 836.859126] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 836.859126] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-b002ed5b-ba60-4692-a7bf-4e5cf941810e"}]} [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] During handling of the above exception, another exception occurred: [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Traceback (most recent call last): [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] with self.rt.instance_claim(context, instance, node, allocs, [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 836.859420] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] return f(*args, **kwargs) [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] self._update(elevated, cn) [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] self._update_to_placement(context, compute_node, startup) [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] return attempt.get(self._wrap_exception) [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] six.reraise(self.value[0], self.value[1], self.value[2]) [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] raise value [ 836.859802] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] self.reportclient.update_from_provider_tree( [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] with catch_all(pd.uuid): [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] self.gen.throw(typ, value, traceback) [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] raise exception.ResourceProviderSyncFailed() [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 836.860305] env[69171]: ERROR nova.compute.manager [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] [ 836.860755] env[69171]: DEBUG nova.compute.utils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 836.861608] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.199s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 836.863109] env[69171]: INFO nova.compute.claims [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 836.866465] env[69171]: DEBUG nova.compute.manager [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Build of instance c5179486-ea0d-4664-8790-9ec0b9eab2ff was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 836.866885] env[69171]: DEBUG nova.compute.manager [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 836.867128] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquiring lock "refresh_cache-c5179486-ea0d-4664-8790-9ec0b9eab2ff" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.867280] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Acquired lock "refresh_cache-c5179486-ea0d-4664-8790-9ec0b9eab2ff" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.867443] env[69171]: DEBUG nova.network.neutron [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 837.036304] env[69171]: INFO nova.compute.manager [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: b189c5ed-d475-44ec-b26e-e7c7014f0a81] Took 1.03 seconds to deallocate network for instance. [ 837.052929] env[69171]: DEBUG nova.compute.manager [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 837.391892] env[69171]: DEBUG nova.network.neutron [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 837.453934] env[69171]: DEBUG nova.network.neutron [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.576867] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.893029] env[69171]: DEBUG nova.scheduler.client.report [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 837.906584] env[69171]: DEBUG nova.scheduler.client.report [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 837.906845] env[69171]: DEBUG nova.compute.provider_tree [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 837.918407] env[69171]: DEBUG nova.scheduler.client.report [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 837.937382] env[69171]: DEBUG nova.scheduler.client.report [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 837.956861] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Releasing lock "refresh_cache-c5179486-ea0d-4664-8790-9ec0b9eab2ff" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.957191] env[69171]: DEBUG nova.compute.manager [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 837.957366] env[69171]: DEBUG nova.compute.manager [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 837.957537] env[69171]: DEBUG nova.network.neutron [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 837.973362] env[69171]: DEBUG nova.network.neutron [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 838.064304] env[69171]: INFO nova.scheduler.client.report [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Deleted allocations for instance b189c5ed-d475-44ec-b26e-e7c7014f0a81 [ 838.225191] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dc09c9e-d735-4819-951a-e79d083291c5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.233975] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b29c637-6e08-4af8-b50f-8de99229ef6a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.266969] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e5f0d7-cf9b-4227-8f50-6d103633bd88 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.275076] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eff20ae0-b063-4436-bf00-72092c1f813e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.288245] env[69171]: DEBUG nova.compute.provider_tree [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 838.477468] env[69171]: DEBUG nova.network.neutron [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.572837] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8cdb513f-cdc3-4a0c-a939-3a1d12bcfde4 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "b189c5ed-d475-44ec-b26e-e7c7014f0a81" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.909s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.810829] env[69171]: ERROR nova.scheduler.client.report [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [req-95f0b690-94c5-4203-9342-87c264276b77] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-95f0b690-94c5-4203-9342-87c264276b77"}]} [ 838.811300] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.950s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.811985] env[69171]: ERROR nova.compute.manager [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 838.811985] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Traceback (most recent call last): [ 838.811985] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 838.811985] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] yield [ 838.811985] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 838.811985] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] self.set_inventory_for_provider( [ 838.811985] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 838.811985] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-95f0b690-94c5-4203-9342-87c264276b77"}]} [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] During handling of the above exception, another exception occurred: [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Traceback (most recent call last): [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] with self.rt.instance_claim(context, instance, node, allocs, [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 838.812333] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] return f(*args, **kwargs) [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] self._update(elevated, cn) [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] self._update_to_placement(context, compute_node, startup) [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] return attempt.get(self._wrap_exception) [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] raise value [ 838.812729] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] self.reportclient.update_from_provider_tree( [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] with catch_all(pd.uuid): [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] self.gen.throw(typ, value, traceback) [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] raise exception.ResourceProviderSyncFailed() [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 838.813208] env[69171]: ERROR nova.compute.manager [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] [ 838.813585] env[69171]: DEBUG nova.compute.utils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 838.814593] env[69171]: DEBUG nova.compute.manager [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Build of instance 36e1bcc1-8503-4786-b3aa-381a110e8bf6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 838.815036] env[69171]: DEBUG nova.compute.manager [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 838.815279] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "refresh_cache-36e1bcc1-8503-4786-b3aa-381a110e8bf6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.815462] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquired lock "refresh_cache-36e1bcc1-8503-4786-b3aa-381a110e8bf6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.815625] env[69171]: DEBUG nova.network.neutron [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 838.817483] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.099s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.818427] env[69171]: INFO nova.compute.claims [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 838.980670] env[69171]: INFO nova.compute.manager [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] [instance: c5179486-ea0d-4664-8790-9ec0b9eab2ff] Took 1.02 seconds to deallocate network for instance. [ 839.077715] env[69171]: DEBUG nova.compute.manager [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 839.340244] env[69171]: DEBUG nova.network.neutron [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 839.413019] env[69171]: DEBUG nova.network.neutron [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.596795] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.843291] env[69171]: DEBUG nova.scheduler.client.report [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 839.856109] env[69171]: DEBUG nova.scheduler.client.report [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 839.856341] env[69171]: DEBUG nova.compute.provider_tree [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 839.868987] env[69171]: DEBUG nova.scheduler.client.report [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 839.884924] env[69171]: DEBUG nova.scheduler.client.report [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 839.916039] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Releasing lock "refresh_cache-36e1bcc1-8503-4786-b3aa-381a110e8bf6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.916286] env[69171]: DEBUG nova.compute.manager [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 839.916465] env[69171]: DEBUG nova.compute.manager [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 839.916635] env[69171]: DEBUG nova.network.neutron [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 839.942975] env[69171]: DEBUG nova.network.neutron [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 840.010275] env[69171]: INFO nova.scheduler.client.report [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Deleted allocations for instance c5179486-ea0d-4664-8790-9ec0b9eab2ff [ 840.193824] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fca2737-e657-49fa-9a7e-e7bed5b23844 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.204765] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225814dd-54c0-4c09-bcd5-62b71968c68e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.257862] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b8a2c4f-0b75-414e-a45d-5b41e77621f6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.269723] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d4a8a3-776a-43d1-987e-237803f66a23 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.291791] env[69171]: DEBUG nova.compute.provider_tree [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 840.450586] env[69171]: DEBUG nova.network.neutron [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 840.521048] env[69171]: DEBUG oslo_concurrency.lockutils [None req-41b98f54-ff8f-4ad1-ad09-050f08e93fd3 tempest-ListImageFiltersTestJSON-1491464896 tempest-ListImageFiltersTestJSON-1491464896-project-member] Lock "c5179486-ea0d-4664-8790-9ec0b9eab2ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.070s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.814924] env[69171]: ERROR nova.scheduler.client.report [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [req-320c573e-cf07-4cbb-b316-053ba3b085fd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-320c573e-cf07-4cbb-b316-053ba3b085fd"}]} [ 840.815315] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.998s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.815904] env[69171]: ERROR nova.compute.manager [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 840.815904] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Traceback (most recent call last): [ 840.815904] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 840.815904] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] yield [ 840.815904] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 840.815904] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] self.set_inventory_for_provider( [ 840.815904] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 840.815904] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-320c573e-cf07-4cbb-b316-053ba3b085fd"}]} [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] During handling of the above exception, another exception occurred: [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Traceback (most recent call last): [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] with self.rt.instance_claim(context, instance, node, allocs, [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 840.816226] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] return f(*args, **kwargs) [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] self._update(elevated, cn) [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] self._update_to_placement(context, compute_node, startup) [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] return attempt.get(self._wrap_exception) [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] raise value [ 840.816556] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] self.reportclient.update_from_provider_tree( [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] with catch_all(pd.uuid): [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] self.gen.throw(typ, value, traceback) [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] raise exception.ResourceProviderSyncFailed() [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 840.816977] env[69171]: ERROR nova.compute.manager [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] [ 840.817360] env[69171]: DEBUG nova.compute.utils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 840.817835] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.837s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.819381] env[69171]: INFO nova.compute.claims [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.822617] env[69171]: DEBUG nova.compute.manager [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Build of instance 8936c08f-2ef1-4b6e-a632-60ca9ad7095d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 840.822755] env[69171]: DEBUG nova.compute.manager [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 840.822862] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquiring lock "refresh_cache-8936c08f-2ef1-4b6e-a632-60ca9ad7095d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.823045] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Acquired lock "refresh_cache-8936c08f-2ef1-4b6e-a632-60ca9ad7095d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.823205] env[69171]: DEBUG nova.network.neutron [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 840.954350] env[69171]: INFO nova.compute.manager [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 36e1bcc1-8503-4786-b3aa-381a110e8bf6] Took 1.04 seconds to deallocate network for instance. [ 841.023707] env[69171]: DEBUG nova.compute.manager [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 841.346212] env[69171]: DEBUG nova.network.neutron [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 841.435917] env[69171]: DEBUG nova.network.neutron [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.544909] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 841.853519] env[69171]: DEBUG nova.scheduler.client.report [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 841.870364] env[69171]: DEBUG nova.scheduler.client.report [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 841.870594] env[69171]: DEBUG nova.compute.provider_tree [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 841.885363] env[69171]: DEBUG nova.scheduler.client.report [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 841.905270] env[69171]: DEBUG nova.scheduler.client.report [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 841.941935] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Releasing lock "refresh_cache-8936c08f-2ef1-4b6e-a632-60ca9ad7095d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.948086] env[69171]: DEBUG nova.compute.manager [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 841.948086] env[69171]: DEBUG nova.compute.manager [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 841.948086] env[69171]: DEBUG nova.network.neutron [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 841.987399] env[69171]: INFO nova.scheduler.client.report [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Deleted allocations for instance 36e1bcc1-8503-4786-b3aa-381a110e8bf6 [ 842.024089] env[69171]: DEBUG nova.network.neutron [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 842.219824] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5059bd90-9991-4a3b-ad98-05eb0f9eb803 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.229310] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7803a6a3-be61-4844-ac41-03c6ca9b331d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.260357] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f36c36b2-373e-49d0-b62e-558ca4713174 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.268543] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-940c077a-ffdb-4e58-902c-b1bf99c46d69 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.282788] env[69171]: DEBUG nova.compute.provider_tree [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 842.498720] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8d9e7446-882d-4aff-87de-f486ed0132ae tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "36e1bcc1-8503-4786-b3aa-381a110e8bf6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.405s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.525477] env[69171]: DEBUG nova.network.neutron [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.810703] env[69171]: ERROR nova.scheduler.client.report [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [req-1e82ab36-656a-4203-9df0-74093e878a81] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1e82ab36-656a-4203-9df0-74093e878a81"}]} [ 842.811123] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.993s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.811821] env[69171]: ERROR nova.compute.manager [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 842.811821] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Traceback (most recent call last): [ 842.811821] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 842.811821] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] yield [ 842.811821] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 842.811821] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] self.set_inventory_for_provider( [ 842.811821] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 842.811821] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1e82ab36-656a-4203-9df0-74093e878a81"}]} [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] During handling of the above exception, another exception occurred: [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Traceback (most recent call last): [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] with self.rt.instance_claim(context, instance, node, allocs, [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 842.812497] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] return f(*args, **kwargs) [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] self._update(elevated, cn) [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] self._update_to_placement(context, compute_node, startup) [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] return attempt.get(self._wrap_exception) [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] six.reraise(self.value[0], self.value[1], self.value[2]) [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] raise value [ 842.812775] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] self.reportclient.update_from_provider_tree( [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] with catch_all(pd.uuid): [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] self.gen.throw(typ, value, traceback) [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] raise exception.ResourceProviderSyncFailed() [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 842.813194] env[69171]: ERROR nova.compute.manager [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] [ 842.813501] env[69171]: DEBUG nova.compute.utils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 842.813954] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.666s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.815749] env[69171]: INFO nova.compute.claims [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.822019] env[69171]: DEBUG nova.compute.manager [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Build of instance 2d5a9f39-c7c1-4aab-91c0-54b51547eab4 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 842.822019] env[69171]: DEBUG nova.compute.manager [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 842.822019] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Acquiring lock "refresh_cache-2d5a9f39-c7c1-4aab-91c0-54b51547eab4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.822019] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Acquired lock "refresh_cache-2d5a9f39-c7c1-4aab-91c0-54b51547eab4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.822247] env[69171]: DEBUG nova.network.neutron [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 843.001440] env[69171]: DEBUG nova.compute.manager [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 843.028268] env[69171]: INFO nova.compute.manager [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] [instance: 8936c08f-2ef1-4b6e-a632-60ca9ad7095d] Took 1.08 seconds to deallocate network for instance. [ 843.349635] env[69171]: DEBUG nova.network.neutron [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 843.465598] env[69171]: DEBUG nova.network.neutron [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.526205] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.851590] env[69171]: DEBUG nova.scheduler.client.report [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 843.865532] env[69171]: DEBUG nova.scheduler.client.report [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 843.865811] env[69171]: DEBUG nova.compute.provider_tree [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 843.879305] env[69171]: DEBUG nova.scheduler.client.report [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 843.899890] env[69171]: DEBUG nova.scheduler.client.report [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 843.971082] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Releasing lock "refresh_cache-2d5a9f39-c7c1-4aab-91c0-54b51547eab4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.971082] env[69171]: DEBUG nova.compute.manager [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 843.971358] env[69171]: DEBUG nova.compute.manager [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 843.971587] env[69171]: DEBUG nova.network.neutron [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 843.987014] env[69171]: DEBUG nova.network.neutron [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 844.066701] env[69171]: INFO nova.scheduler.client.report [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Deleted allocations for instance 8936c08f-2ef1-4b6e-a632-60ca9ad7095d [ 844.206640] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b76e520a-1185-4b9b-bbe6-aadd5b6b581b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.215297] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ee58d57-adc1-40fd-9bed-0c69df2cf68f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.245814] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca3ca837-9e58-4002-a950-aec0c56d7c08 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.254331] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da1fedfc-5c12-4ffd-ada8-d66a055b9525 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.269833] env[69171]: DEBUG nova.compute.provider_tree [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 844.489503] env[69171]: DEBUG nova.network.neutron [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.585347] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0795060b-b8af-4212-932c-46961f9737c6 tempest-ListServerFiltersTestJSON-2020766152 tempest-ListServerFiltersTestJSON-2020766152-project-member] Lock "8936c08f-2ef1-4b6e-a632-60ca9ad7095d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.888s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.791940] env[69171]: ERROR nova.scheduler.client.report [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [req-0ccda1e2-f230-4e51-b48b-9cf7f9901917] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0ccda1e2-f230-4e51-b48b-9cf7f9901917"}]} [ 844.792337] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.978s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.793051] env[69171]: ERROR nova.compute.manager [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 844.793051] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Traceback (most recent call last): [ 844.793051] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 844.793051] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] yield [ 844.793051] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 844.793051] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] self.set_inventory_for_provider( [ 844.793051] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 844.793051] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-0ccda1e2-f230-4e51-b48b-9cf7f9901917"}]} [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] During handling of the above exception, another exception occurred: [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Traceback (most recent call last): [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] with self.rt.instance_claim(context, instance, node, allocs, [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 844.793368] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] return f(*args, **kwargs) [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] self._update(elevated, cn) [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] self._update_to_placement(context, compute_node, startup) [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] return attempt.get(self._wrap_exception) [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] raise value [ 844.793640] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] self.reportclient.update_from_provider_tree( [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] with catch_all(pd.uuid): [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] self.gen.throw(typ, value, traceback) [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] raise exception.ResourceProviderSyncFailed() [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 844.793990] env[69171]: ERROR nova.compute.manager [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] [ 844.794309] env[69171]: DEBUG nova.compute.utils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 844.794861] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.448s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.796412] env[69171]: INFO nova.compute.claims [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 844.800317] env[69171]: DEBUG nova.compute.manager [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Build of instance 8e90508a-fc26-4e9d-ac70-054b249bc2a3 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 844.800761] env[69171]: DEBUG nova.compute.manager [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 844.800988] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Acquiring lock "refresh_cache-8e90508a-fc26-4e9d-ac70-054b249bc2a3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.801149] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Acquired lock "refresh_cache-8e90508a-fc26-4e9d-ac70-054b249bc2a3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.801308] env[69171]: DEBUG nova.network.neutron [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 844.995054] env[69171]: INFO nova.compute.manager [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] [instance: 2d5a9f39-c7c1-4aab-91c0-54b51547eab4] Took 1.02 seconds to deallocate network for instance. [ 845.090071] env[69171]: DEBUG nova.compute.manager [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 845.323233] env[69171]: DEBUG nova.network.neutron [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 845.424905] env[69171]: DEBUG nova.network.neutron [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.609291] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.687778] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Acquiring lock "5001d2de-7fdd-479c-9a10-8b3a3e5862b6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.688035] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Lock "5001d2de-7fdd-479c-9a10-8b3a3e5862b6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.826746] env[69171]: DEBUG nova.scheduler.client.report [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 845.842565] env[69171]: DEBUG nova.scheduler.client.report [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 845.842813] env[69171]: DEBUG nova.compute.provider_tree [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 845.854434] env[69171]: DEBUG nova.scheduler.client.report [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 845.874153] env[69171]: DEBUG nova.scheduler.client.report [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 845.927327] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Releasing lock "refresh_cache-8e90508a-fc26-4e9d-ac70-054b249bc2a3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.927596] env[69171]: DEBUG nova.compute.manager [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 845.928191] env[69171]: DEBUG nova.compute.manager [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 845.928191] env[69171]: DEBUG nova.network.neutron [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 845.947389] env[69171]: DEBUG nova.network.neutron [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 846.022399] env[69171]: INFO nova.scheduler.client.report [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Deleted allocations for instance 2d5a9f39-c7c1-4aab-91c0-54b51547eab4 [ 846.177656] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffb8fd7-66fb-463e-b465-f3f5b9b8382e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.186330] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95a1b932-6888-4802-a264-a658b35daf1a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.226219] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27912ca9-a666-4ce1-a307-a30ff9d78ab6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.235194] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26873a29-5384-4aaf-9668-547ffd023828 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.251349] env[69171]: DEBUG nova.compute.provider_tree [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 846.453106] env[69171]: DEBUG nova.network.neutron [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 846.533725] env[69171]: DEBUG oslo_concurrency.lockutils [None req-606562ab-5081-4d88-8f94-ebf2a801de80 tempest-ServerActionsTestOtherB-1613971756 tempest-ServerActionsTestOtherB-1613971756-project-member] Lock "2d5a9f39-c7c1-4aab-91c0-54b51547eab4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 52.344s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.775262] env[69171]: ERROR nova.scheduler.client.report [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [req-4e554b6b-1c39-421d-b2ae-6a6f61e26b17] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-4e554b6b-1c39-421d-b2ae-6a6f61e26b17"}]} [ 846.776127] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.981s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.776793] env[69171]: ERROR nova.compute.manager [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 846.776793] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Traceback (most recent call last): [ 846.776793] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 846.776793] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] yield [ 846.776793] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 846.776793] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] self.set_inventory_for_provider( [ 846.776793] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 846.776793] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-4e554b6b-1c39-421d-b2ae-6a6f61e26b17"}]} [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] During handling of the above exception, another exception occurred: [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Traceback (most recent call last): [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] with self.rt.instance_claim(context, instance, node, allocs, [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 846.777094] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] return f(*args, **kwargs) [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] self._update(elevated, cn) [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] self._update_to_placement(context, compute_node, startup) [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] return attempt.get(self._wrap_exception) [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] six.reraise(self.value[0], self.value[1], self.value[2]) [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] raise value [ 846.777466] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] self.reportclient.update_from_provider_tree( [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] with catch_all(pd.uuid): [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] self.gen.throw(typ, value, traceback) [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] raise exception.ResourceProviderSyncFailed() [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 846.777877] env[69171]: ERROR nova.compute.manager [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] [ 846.778242] env[69171]: DEBUG nova.compute.utils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 846.778943] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.314s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.780506] env[69171]: INFO nova.compute.claims [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 846.785690] env[69171]: DEBUG nova.compute.manager [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Build of instance a0941eb1-4a98-450b-8318-e8f60db1ac03 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 846.785690] env[69171]: DEBUG nova.compute.manager [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 846.785690] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquiring lock "refresh_cache-a0941eb1-4a98-450b-8318-e8f60db1ac03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.785690] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquired lock "refresh_cache-a0941eb1-4a98-450b-8318-e8f60db1ac03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.786359] env[69171]: DEBUG nova.network.neutron [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 846.955995] env[69171]: INFO nova.compute.manager [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] [instance: 8e90508a-fc26-4e9d-ac70-054b249bc2a3] Took 1.03 seconds to deallocate network for instance. [ 847.037326] env[69171]: DEBUG nova.compute.manager [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 847.312547] env[69171]: DEBUG nova.network.neutron [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 847.435535] env[69171]: DEBUG nova.network.neutron [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.565756] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.811046] env[69171]: DEBUG nova.scheduler.client.report [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 847.825333] env[69171]: DEBUG nova.scheduler.client.report [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 847.825556] env[69171]: DEBUG nova.compute.provider_tree [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 847.837270] env[69171]: DEBUG nova.scheduler.client.report [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 847.857489] env[69171]: DEBUG nova.scheduler.client.report [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 847.942405] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Releasing lock "refresh_cache-a0941eb1-4a98-450b-8318-e8f60db1ac03" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.942671] env[69171]: DEBUG nova.compute.manager [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 847.942852] env[69171]: DEBUG nova.compute.manager [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 847.943030] env[69171]: DEBUG nova.network.neutron [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 847.963876] env[69171]: DEBUG nova.network.neutron [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 847.988985] env[69171]: INFO nova.scheduler.client.report [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Deleted allocations for instance 8e90508a-fc26-4e9d-ac70-054b249bc2a3 [ 848.104215] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4022ff89-7cc4-49e9-ac6d-28cb3f5b2b94 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.112972] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13e9c216-c3b2-461e-bf6f-378ce5bfffec {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.142979] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8632e9a-a232-4c5c-bfea-105dc0160965 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.151217] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d0f043-132b-4ff1-a9e6-0c3cbb24efd5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.165298] env[69171]: DEBUG nova.compute.provider_tree [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 848.467581] env[69171]: DEBUG nova.network.neutron [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.497196] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3ceb1e74-b4bc-4570-bbd8-c014865dd4bb tempest-AttachInterfacesV270Test-1612201928 tempest-AttachInterfacesV270Test-1612201928-project-member] Lock "8e90508a-fc26-4e9d-ac70-054b249bc2a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.621s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.687101] env[69171]: ERROR nova.scheduler.client.report [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [req-a17b77c1-a724-45d7-b0a2-443329500566] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a17b77c1-a724-45d7-b0a2-443329500566"}]} [ 848.687554] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.909s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 848.688162] env[69171]: ERROR nova.compute.manager [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 848.688162] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Traceback (most recent call last): [ 848.688162] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 848.688162] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] yield [ 848.688162] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 848.688162] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] self.set_inventory_for_provider( [ 848.688162] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 848.688162] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a17b77c1-a724-45d7-b0a2-443329500566"}]} [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] During handling of the above exception, another exception occurred: [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Traceback (most recent call last): [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] with self.rt.instance_claim(context, instance, node, allocs, [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 848.688437] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] return f(*args, **kwargs) [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] self._update(elevated, cn) [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] self._update_to_placement(context, compute_node, startup) [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] return attempt.get(self._wrap_exception) [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] six.reraise(self.value[0], self.value[1], self.value[2]) [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] raise value [ 848.688766] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] self.reportclient.update_from_provider_tree( [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] with catch_all(pd.uuid): [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] self.gen.throw(typ, value, traceback) [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] raise exception.ResourceProviderSyncFailed() [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 848.689209] env[69171]: ERROR nova.compute.manager [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] [ 848.689596] env[69171]: DEBUG nova.compute.utils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 848.690756] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.224s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.692250] env[69171]: INFO nova.compute.claims [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 848.695636] env[69171]: DEBUG nova.compute.manager [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Build of instance ecec698a-7570-4400-ad8c-9215cd6733ef was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 848.696068] env[69171]: DEBUG nova.compute.manager [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 848.696300] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "refresh_cache-ecec698a-7570-4400-ad8c-9215cd6733ef" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.696445] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquired lock "refresh_cache-ecec698a-7570-4400-ad8c-9215cd6733ef" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.696606] env[69171]: DEBUG nova.network.neutron [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 848.970819] env[69171]: INFO nova.compute.manager [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: a0941eb1-4a98-450b-8318-e8f60db1ac03] Took 1.03 seconds to deallocate network for instance. [ 849.000586] env[69171]: DEBUG nova.compute.manager [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 849.220124] env[69171]: DEBUG nova.network.neutron [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 849.359578] env[69171]: DEBUG nova.network.neutron [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.529857] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.727104] env[69171]: DEBUG nova.scheduler.client.report [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 849.749133] env[69171]: DEBUG nova.scheduler.client.report [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 849.749342] env[69171]: DEBUG nova.compute.provider_tree [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 849.767966] env[69171]: DEBUG nova.scheduler.client.report [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 849.796967] env[69171]: DEBUG nova.scheduler.client.report [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 849.862976] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Releasing lock "refresh_cache-ecec698a-7570-4400-ad8c-9215cd6733ef" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.862976] env[69171]: DEBUG nova.compute.manager [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 849.862976] env[69171]: DEBUG nova.compute.manager [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 849.862976] env[69171]: DEBUG nova.network.neutron [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 849.880752] env[69171]: DEBUG nova.network.neutron [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 849.999563] env[69171]: INFO nova.scheduler.client.report [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Deleted allocations for instance a0941eb1-4a98-450b-8318-e8f60db1ac03 [ 850.062908] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7698f588-0484-4598-b2d4-93c1de707de9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.070513] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-407237b5-8358-4ca7-ab08-2122c185bb46 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.104258] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bb4e954-6688-4ddf-86fa-60366894c862 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.112477] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8766f9d5-d8e7-42d3-b833-974ca8805b8b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.126565] env[69171]: DEBUG nova.compute.provider_tree [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 850.387480] env[69171]: DEBUG nova.network.neutron [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.511728] env[69171]: DEBUG oslo_concurrency.lockutils [None req-65cec10c-5dca-4c95-870c-fb248b53c91e tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "a0941eb1-4a98-450b-8318-e8f60db1ac03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.058s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.648126] env[69171]: ERROR nova.scheduler.client.report [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [req-57a28b4b-97e2-4ab7-ad80-7542ecbb47f9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-57a28b4b-97e2-4ab7-ad80-7542ecbb47f9"}]} [ 850.648499] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.958s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.649476] env[69171]: ERROR nova.compute.manager [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 850.649476] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Traceback (most recent call last): [ 850.649476] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 850.649476] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] yield [ 850.649476] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 850.649476] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] self.set_inventory_for_provider( [ 850.649476] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 850.649476] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-57a28b4b-97e2-4ab7-ad80-7542ecbb47f9"}]} [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] During handling of the above exception, another exception occurred: [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Traceback (most recent call last): [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] with self.rt.instance_claim(context, instance, node, allocs, [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 850.649863] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] return f(*args, **kwargs) [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] self._update(elevated, cn) [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] self._update_to_placement(context, compute_node, startup) [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] return attempt.get(self._wrap_exception) [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] six.reraise(self.value[0], self.value[1], self.value[2]) [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] raise value [ 850.650199] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] self.reportclient.update_from_provider_tree( [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] with catch_all(pd.uuid): [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] self.gen.throw(typ, value, traceback) [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] raise exception.ResourceProviderSyncFailed() [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 850.650618] env[69171]: ERROR nova.compute.manager [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] [ 850.650967] env[69171]: DEBUG nova.compute.utils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 850.651519] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.075s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 850.652987] env[69171]: INFO nova.compute.claims [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 850.655838] env[69171]: DEBUG nova.compute.manager [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Build of instance a74e930a-9871-4aa6-8997-e5ef6042c9d3 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 850.656153] env[69171]: DEBUG nova.compute.manager [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 850.656384] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Acquiring lock "refresh_cache-a74e930a-9871-4aa6-8997-e5ef6042c9d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.656533] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Acquired lock "refresh_cache-a74e930a-9871-4aa6-8997-e5ef6042c9d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.656693] env[69171]: DEBUG nova.network.neutron [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 850.891989] env[69171]: INFO nova.compute.manager [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: ecec698a-7570-4400-ad8c-9215cd6733ef] Took 1.03 seconds to deallocate network for instance. [ 851.016352] env[69171]: DEBUG nova.compute.manager [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 851.188257] env[69171]: DEBUG nova.network.neutron [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 851.266665] env[69171]: DEBUG nova.network.neutron [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.539787] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.692421] env[69171]: DEBUG nova.scheduler.client.report [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 851.708392] env[69171]: DEBUG nova.scheduler.client.report [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 851.708546] env[69171]: DEBUG nova.compute.provider_tree [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 851.729480] env[69171]: DEBUG nova.scheduler.client.report [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 851.752876] env[69171]: DEBUG nova.scheduler.client.report [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 851.768273] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Releasing lock "refresh_cache-a74e930a-9871-4aa6-8997-e5ef6042c9d3" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.768534] env[69171]: DEBUG nova.compute.manager [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 851.768775] env[69171]: DEBUG nova.compute.manager [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] [instance: a74e930a-9871-4aa6-8997-e5ef6042c9d3] Skipping network deallocation for instance since networking was not requested. {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2286}} [ 851.926531] env[69171]: INFO nova.scheduler.client.report [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Deleted allocations for instance ecec698a-7570-4400-ad8c-9215cd6733ef [ 852.022096] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a26d21-c59a-472e-b040-640cae462844 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.030068] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd6ef08-263a-42b8-97a9-3a13828e9316 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.064387] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e18508ed-9b78-4310-91ff-c9be7fd39594 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.073446] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01b7ba82-797f-4cd1-87bb-563a995365af {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.088603] env[69171]: DEBUG nova.compute.provider_tree [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 852.435190] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ed0764ce-9cbc-44cb-aa3a-a22f176fa052 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "ecec698a-7570-4400-ad8c-9215cd6733ef" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.952s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.620237] env[69171]: ERROR nova.scheduler.client.report [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [req-a00cf2ee-32de-41a5-b6fb-71081b0d4b6c] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a00cf2ee-32de-41a5-b6fb-71081b0d4b6c"}]} [ 852.620615] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.969s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.621212] env[69171]: ERROR nova.compute.manager [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 852.621212] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Traceback (most recent call last): [ 852.621212] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 852.621212] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] yield [ 852.621212] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 852.621212] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] self.set_inventory_for_provider( [ 852.621212] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 852.621212] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a00cf2ee-32de-41a5-b6fb-71081b0d4b6c"}]} [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] During handling of the above exception, another exception occurred: [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Traceback (most recent call last): [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] with self.rt.instance_claim(context, instance, node, allocs, [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 852.622155] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] return f(*args, **kwargs) [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] self._update(elevated, cn) [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] self._update_to_placement(context, compute_node, startup) [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] return attempt.get(self._wrap_exception) [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] raise value [ 852.622763] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] self.reportclient.update_from_provider_tree( [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] with catch_all(pd.uuid): [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] self.gen.throw(typ, value, traceback) [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] raise exception.ResourceProviderSyncFailed() [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 852.623365] env[69171]: ERROR nova.compute.manager [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] [ 852.623861] env[69171]: DEBUG nova.compute.utils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 852.624866] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.028s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.629143] env[69171]: INFO nova.compute.claims [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 852.631083] env[69171]: DEBUG nova.compute.manager [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Build of instance d950fb6b-7977-40c1-b9e1-a89bc2e9fec6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 852.632241] env[69171]: DEBUG nova.compute.manager [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 852.632786] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Acquiring lock "refresh_cache-d950fb6b-7977-40c1-b9e1-a89bc2e9fec6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.633172] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Acquired lock "refresh_cache-d950fb6b-7977-40c1-b9e1-a89bc2e9fec6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.634626] env[69171]: DEBUG nova.network.neutron [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 852.807605] env[69171]: INFO nova.scheduler.client.report [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Deleted allocations for instance a74e930a-9871-4aa6-8997-e5ef6042c9d3 [ 852.937890] env[69171]: DEBUG nova.compute.manager [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 853.184698] env[69171]: DEBUG nova.network.neutron [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 853.280127] env[69171]: DEBUG nova.network.neutron [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 853.315049] env[69171]: DEBUG oslo_concurrency.lockutils [None req-27f76e86-ab2d-438c-8107-1bfc87b02904 tempest-ServerShowV254Test-672385691 tempest-ServerShowV254Test-672385691-project-member] Lock "a74e930a-9871-4aa6-8997-e5ef6042c9d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.890s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.467254] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.682764] env[69171]: DEBUG nova.scheduler.client.report [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 853.696888] env[69171]: DEBUG nova.scheduler.client.report [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 853.697193] env[69171]: DEBUG nova.compute.provider_tree [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 853.714435] env[69171]: DEBUG nova.scheduler.client.report [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 853.734619] env[69171]: DEBUG nova.scheduler.client.report [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 853.788635] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Releasing lock "refresh_cache-d950fb6b-7977-40c1-b9e1-a89bc2e9fec6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 853.789282] env[69171]: DEBUG nova.compute.manager [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 853.790499] env[69171]: DEBUG nova.compute.manager [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] [instance: d950fb6b-7977-40c1-b9e1-a89bc2e9fec6] Skipping network deallocation for instance since networking was not requested. {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2286}} [ 853.821593] env[69171]: DEBUG nova.compute.manager [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 853.882540] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquiring lock "3cc37f29-ae73-4a6a-8bb1-88484c59d83c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.882807] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "3cc37f29-ae73-4a6a-8bb1-88484c59d83c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.999470] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-396af8b2-6b3a-4dcd-96f9-16c067ad35ca {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.009111] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03511bae-baae-45fa-b2b3-8d38d9cd3d9c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.044748] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3abaf08-3220-4a0d-95fc-dddb5bfc0b02 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.053513] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ffb882a-0085-410d-a54e-db7030d4bb8e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.069564] env[69171]: DEBUG nova.compute.provider_tree [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 854.349224] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.605767] env[69171]: ERROR nova.scheduler.client.report [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [req-5a1f1d20-c9dc-4ee1-8d87-28b05fdfb7fd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5a1f1d20-c9dc-4ee1-8d87-28b05fdfb7fd"}]} [ 854.606523] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.982s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.607180] env[69171]: ERROR nova.compute.manager [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 854.607180] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Traceback (most recent call last): [ 854.607180] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 854.607180] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] yield [ 854.607180] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 854.607180] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] self.set_inventory_for_provider( [ 854.607180] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 854.607180] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5a1f1d20-c9dc-4ee1-8d87-28b05fdfb7fd"}]} [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] During handling of the above exception, another exception occurred: [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Traceback (most recent call last): [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] with self.rt.instance_claim(context, instance, node, allocs, [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 854.607624] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] return f(*args, **kwargs) [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] self._update(elevated, cn) [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] self._update_to_placement(context, compute_node, startup) [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] return attempt.get(self._wrap_exception) [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] raise value [ 854.608488] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] self.reportclient.update_from_provider_tree( [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] with catch_all(pd.uuid): [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] self.gen.throw(typ, value, traceback) [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] raise exception.ResourceProviderSyncFailed() [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 854.609182] env[69171]: ERROR nova.compute.manager [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] [ 854.610056] env[69171]: DEBUG nova.compute.utils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 854.610056] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.065s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.611289] env[69171]: INFO nova.compute.claims [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.617167] env[69171]: DEBUG nova.compute.manager [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Build of instance 13d5116e-46ab-40d4-a507-3d34fedd7d4d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 854.617167] env[69171]: DEBUG nova.compute.manager [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 854.617167] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Acquiring lock "refresh_cache-13d5116e-46ab-40d4-a507-3d34fedd7d4d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.617167] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Acquired lock "refresh_cache-13d5116e-46ab-40d4-a507-3d34fedd7d4d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.617345] env[69171]: DEBUG nova.network.neutron [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 854.650154] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "a8c74dc9-6fb9-46f1-b558-2797ccfdfd16" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.650382] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "a8c74dc9-6fb9-46f1-b558-2797ccfdfd16" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.827527] env[69171]: INFO nova.scheduler.client.report [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Deleted allocations for instance d950fb6b-7977-40c1-b9e1-a89bc2e9fec6 [ 855.138153] env[69171]: DEBUG nova.network.neutron [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.254645] env[69171]: DEBUG nova.network.neutron [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.335507] env[69171]: DEBUG oslo_concurrency.lockutils [None req-bd94fe3e-706d-4d7e-93a3-01be4d047502 tempest-ServerShowV257Test-295547851 tempest-ServerShowV257Test-295547851-project-member] Lock "d950fb6b-7977-40c1-b9e1-a89bc2e9fec6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.915s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.638976] env[69171]: DEBUG nova.scheduler.client.report [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 855.654606] env[69171]: DEBUG nova.scheduler.client.report [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 855.654847] env[69171]: DEBUG nova.compute.provider_tree [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 855.670529] env[69171]: DEBUG nova.scheduler.client.report [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 855.689439] env[69171]: DEBUG nova.scheduler.client.report [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 855.744092] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "87abf9b1-06f4-412f-bb1d-f1e4cb77ef82" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.744389] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "87abf9b1-06f4-412f-bb1d-f1e4cb77ef82" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.757912] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Releasing lock "refresh_cache-13d5116e-46ab-40d4-a507-3d34fedd7d4d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.757912] env[69171]: DEBUG nova.compute.manager [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 855.758088] env[69171]: DEBUG nova.compute.manager [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 855.758890] env[69171]: DEBUG nova.network.neutron [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 855.775260] env[69171]: DEBUG nova.network.neutron [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 855.841946] env[69171]: DEBUG nova.compute.manager [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 856.077844] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7bd5c9f-9904-4547-9e6b-f326ec18e49e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.086094] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213bce45-02f9-40f5-9de7-b0f4a69c8f43 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.116568] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f304c28d-ebcf-4f3f-88cd-28442cf4ae88 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.124715] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c440e9c-8eed-48f1-927a-4d30d25e21b8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.138485] env[69171]: DEBUG nova.compute.provider_tree [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 856.278117] env[69171]: DEBUG nova.network.neutron [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 856.365081] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.659135] env[69171]: ERROR nova.scheduler.client.report [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [req-f7452c63-8385-4249-a7ab-d4fa86b5ec74] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f7452c63-8385-4249-a7ab-d4fa86b5ec74"}]} [ 856.659530] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.050s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.660231] env[69171]: ERROR nova.compute.manager [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 856.660231] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Traceback (most recent call last): [ 856.660231] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 856.660231] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] yield [ 856.660231] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 856.660231] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] self.set_inventory_for_provider( [ 856.660231] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 856.660231] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f7452c63-8385-4249-a7ab-d4fa86b5ec74"}]} [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] During handling of the above exception, another exception occurred: [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Traceback (most recent call last): [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] with self.rt.instance_claim(context, instance, node, allocs, [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 856.660526] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] return f(*args, **kwargs) [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] self._update(elevated, cn) [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] self._update_to_placement(context, compute_node, startup) [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] return attempt.get(self._wrap_exception) [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] six.reraise(self.value[0], self.value[1], self.value[2]) [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] raise value [ 856.660817] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] self.reportclient.update_from_provider_tree( [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] with catch_all(pd.uuid): [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] self.gen.throw(typ, value, traceback) [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] raise exception.ResourceProviderSyncFailed() [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 856.661328] env[69171]: ERROR nova.compute.manager [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] [ 856.661696] env[69171]: DEBUG nova.compute.utils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 856.662311] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.136s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.664141] env[69171]: INFO nova.compute.claims [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 856.666731] env[69171]: DEBUG nova.compute.manager [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Build of instance 0da1037f-5468-4926-be0a-8a91f31f8bdb was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 856.667149] env[69171]: DEBUG nova.compute.manager [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 856.667374] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquiring lock "refresh_cache-0da1037f-5468-4926-be0a-8a91f31f8bdb" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.667522] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquired lock "refresh_cache-0da1037f-5468-4926-be0a-8a91f31f8bdb" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.667681] env[69171]: DEBUG nova.network.neutron [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 856.781255] env[69171]: INFO nova.compute.manager [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] [instance: 13d5116e-46ab-40d4-a507-3d34fedd7d4d] Took 1.02 seconds to deallocate network for instance. [ 857.020248] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.020508] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.190034] env[69171]: DEBUG nova.network.neutron [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 857.252408] env[69171]: DEBUG nova.network.neutron [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.528032] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 857.528032] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Starting heal instance info cache {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9929}} [ 857.690013] env[69171]: DEBUG nova.scheduler.client.report [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 857.703893] env[69171]: DEBUG nova.scheduler.client.report [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 857.704149] env[69171]: DEBUG nova.compute.provider_tree [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 857.715133] env[69171]: DEBUG nova.scheduler.client.report [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 857.732780] env[69171]: DEBUG nova.scheduler.client.report [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 857.755421] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Releasing lock "refresh_cache-0da1037f-5468-4926-be0a-8a91f31f8bdb" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.755649] env[69171]: DEBUG nova.compute.manager [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 857.755843] env[69171]: DEBUG nova.compute.manager [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 0da1037f-5468-4926-be0a-8a91f31f8bdb] Skipping network deallocation for instance since networking was not requested. {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2286}} [ 857.807681] env[69171]: INFO nova.scheduler.client.report [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Deleted allocations for instance 13d5116e-46ab-40d4-a507-3d34fedd7d4d [ 857.956824] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ca6dc00-0d3b-4674-97ab-8bafdc8a24a9 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.965186] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab745eb9-e2e1-482c-9905-3760ef51476d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.995587] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a68792b7-ffdf-4b91-b208-8538df15ddaf {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.004021] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e180a573-993a-4fe7-b8ac-c6c455b868a7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.019044] env[69171]: DEBUG nova.compute.provider_tree [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 858.046747] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.046904] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquired lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.047065] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Forcefully refreshing network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 858.318295] env[69171]: DEBUG oslo_concurrency.lockutils [None req-94f7dd00-dc60-49bf-b67e-634720a57584 tempest-ServerActionsTestOtherA-1886098112 tempest-ServerActionsTestOtherA-1886098112-project-member] Lock "13d5116e-46ab-40d4-a507-3d34fedd7d4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.607s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.542129] env[69171]: ERROR nova.scheduler.client.report [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [req-03ebdf1d-83f8-4b39-9337-71135dd905c1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-03ebdf1d-83f8-4b39-9337-71135dd905c1"}]} [ 858.542648] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.880s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 858.543343] env[69171]: ERROR nova.compute.manager [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 858.543343] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Traceback (most recent call last): [ 858.543343] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 858.543343] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] yield [ 858.543343] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 858.543343] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] self.set_inventory_for_provider( [ 858.543343] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 858.543343] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-03ebdf1d-83f8-4b39-9337-71135dd905c1"}]} [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] During handling of the above exception, another exception occurred: [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Traceback (most recent call last): [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] with self.rt.instance_claim(context, instance, node, allocs, [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 858.543750] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] return f(*args, **kwargs) [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] self._update(elevated, cn) [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] self._update_to_placement(context, compute_node, startup) [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] return attempt.get(self._wrap_exception) [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] six.reraise(self.value[0], self.value[1], self.value[2]) [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] raise value [ 858.544238] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] self.reportclient.update_from_provider_tree( [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] with catch_all(pd.uuid): [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] self.gen.throw(typ, value, traceback) [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] raise exception.ResourceProviderSyncFailed() [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 858.544826] env[69171]: ERROR nova.compute.manager [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] [ 858.545302] env[69171]: DEBUG nova.compute.utils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 858.545398] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.936s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 858.546856] env[69171]: INFO nova.compute.claims [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 858.549560] env[69171]: DEBUG nova.compute.manager [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Build of instance 87cec11c-66c7-4b71-946b-1e8e7d72f9aa was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 858.549985] env[69171]: DEBUG nova.compute.manager [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 858.550254] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquiring lock "refresh_cache-87cec11c-66c7-4b71-946b-1e8e7d72f9aa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.550408] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Acquired lock "refresh_cache-87cec11c-66c7-4b71-946b-1e8e7d72f9aa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.550571] env[69171]: DEBUG nova.network.neutron [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 858.568309] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 858.789624] env[69171]: INFO nova.scheduler.client.report [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Deleted allocations for instance 0da1037f-5468-4926-be0a-8a91f31f8bdb [ 858.821942] env[69171]: DEBUG nova.compute.manager [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 859.077063] env[69171]: DEBUG nova.network.neutron [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 859.132615] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.172070] env[69171]: DEBUG nova.network.neutron [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.298338] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a0f95229-59ce-48b2-9b96-1d3bd4f6c22a tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "0da1037f-5468-4926-be0a-8a91f31f8bdb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.095s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.346218] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 859.578349] env[69171]: DEBUG nova.scheduler.client.report [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 859.593526] env[69171]: DEBUG nova.scheduler.client.report [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 859.594032] env[69171]: DEBUG nova.compute.provider_tree [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 859.608315] env[69171]: DEBUG nova.scheduler.client.report [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 859.634914] env[69171]: DEBUG nova.scheduler.client.report [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 859.637218] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Releasing lock "refresh_cache-f5a71cca-1e32-4c3c-9bb8-83e634a864f7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.637411] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f5a71cca-1e32-4c3c-9bb8-83e634a864f7] Updated the network info_cache for instance {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10000}} [ 859.638624] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.638624] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.638624] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.638624] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.638624] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.638624] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.638927] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=69171) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10548}} [ 859.638927] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager.update_available_resource {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 859.675633] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Releasing lock "refresh_cache-87cec11c-66c7-4b71-946b-1e8e7d72f9aa" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.675633] env[69171]: DEBUG nova.compute.manager [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 859.675633] env[69171]: DEBUG nova.compute.manager [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] [instance: 87cec11c-66c7-4b71-946b-1e8e7d72f9aa] Skipping network deallocation for instance since networking was not requested. {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2286}} [ 859.804286] env[69171]: DEBUG nova.compute.manager [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 859.899177] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664d73e3-e305-46b5-84fc-594e0819d74d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.909068] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d62c363-4288-46b1-9507-85f32f1a943e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.941669] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c77d93e-4c0b-478f-9248-043f0fbfed5d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.950058] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a94082ca-0aa5-466d-9341-b81ac511a075 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.965875] env[69171]: DEBUG nova.compute.provider_tree [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 860.142938] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.330848] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.489140] env[69171]: ERROR nova.scheduler.client.report [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [req-7aafd9b8-07d5-449e-bc92-6fed541f395d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-7aafd9b8-07d5-449e-bc92-6fed541f395d"}]} [ 860.489576] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.944s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.490272] env[69171]: ERROR nova.compute.manager [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 860.490272] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Traceback (most recent call last): [ 860.490272] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 860.490272] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] yield [ 860.490272] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 860.490272] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] self.set_inventory_for_provider( [ 860.490272] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 860.490272] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-7aafd9b8-07d5-449e-bc92-6fed541f395d"}]} [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] During handling of the above exception, another exception occurred: [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Traceback (most recent call last): [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] with self.rt.instance_claim(context, instance, node, allocs, [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 860.490517] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] return f(*args, **kwargs) [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] self._update(elevated, cn) [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] self._update_to_placement(context, compute_node, startup) [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] return attempt.get(self._wrap_exception) [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] raise value [ 860.490804] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] self.reportclient.update_from_provider_tree( [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] with catch_all(pd.uuid): [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] self.gen.throw(typ, value, traceback) [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] raise exception.ResourceProviderSyncFailed() [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 860.491184] env[69171]: ERROR nova.compute.manager [instance: 6071a99a-ff47-4161-8a1e-65385318891d] [ 860.491613] env[69171]: DEBUG nova.compute.utils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 860.492434] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.927s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.494278] env[69171]: INFO nova.compute.claims [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.496965] env[69171]: DEBUG nova.compute.manager [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Build of instance 6071a99a-ff47-4161-8a1e-65385318891d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 860.497430] env[69171]: DEBUG nova.compute.manager [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 860.497619] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Acquiring lock "refresh_cache-6071a99a-ff47-4161-8a1e-65385318891d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.497830] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Acquired lock "refresh_cache-6071a99a-ff47-4161-8a1e-65385318891d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.497911] env[69171]: DEBUG nova.network.neutron [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 860.701428] env[69171]: INFO nova.scheduler.client.report [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Deleted allocations for instance 87cec11c-66c7-4b71-946b-1e8e7d72f9aa [ 861.023673] env[69171]: DEBUG nova.network.neutron [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.105042] env[69171]: DEBUG nova.network.neutron [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.208695] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5d39c829-c3cd-4ee1-bd2c-4bb252805a76 tempest-ServerShowV247Test-796349253 tempest-ServerShowV247Test-796349253-project-member] Lock "87cec11c-66c7-4b71-946b-1e8e7d72f9aa" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.321s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.524280] env[69171]: DEBUG nova.scheduler.client.report [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 861.540922] env[69171]: DEBUG nova.scheduler.client.report [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 861.541161] env[69171]: DEBUG nova.compute.provider_tree [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 861.552758] env[69171]: DEBUG nova.scheduler.client.report [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 861.574743] env[69171]: DEBUG nova.scheduler.client.report [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 861.608768] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Releasing lock "refresh_cache-6071a99a-ff47-4161-8a1e-65385318891d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 861.610244] env[69171]: DEBUG nova.compute.manager [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 861.610244] env[69171]: DEBUG nova.compute.manager [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 861.610244] env[69171]: DEBUG nova.network.neutron [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 861.634752] env[69171]: DEBUG nova.network.neutron [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 861.712093] env[69171]: DEBUG nova.compute.manager [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 861.776230] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e3d71f6-c9b2-456b-8bb7-bac37409bfdc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.784829] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1aff5a2f-af11-4e5b-9bf4-f3504202691e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.815020] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbb0d37e-e05d-4cec-a374-1e6de2e1cffa {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.823594] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100c0744-bf2a-4635-b134-e6e477b554c7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.837935] env[69171]: DEBUG nova.compute.provider_tree [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 862.139567] env[69171]: DEBUG nova.network.neutron [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.237032] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.358665] env[69171]: ERROR nova.scheduler.client.report [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [req-62f3bb6d-a322-4289-8057-b6efa93f53d9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-62f3bb6d-a322-4289-8057-b6efa93f53d9"}]} [ 862.359085] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.867s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.359799] env[69171]: ERROR nova.compute.manager [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 862.359799] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Traceback (most recent call last): [ 862.359799] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 862.359799] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] yield [ 862.359799] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 862.359799] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] self.set_inventory_for_provider( [ 862.359799] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 862.359799] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-62f3bb6d-a322-4289-8057-b6efa93f53d9"}]} [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] During handling of the above exception, another exception occurred: [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Traceback (most recent call last): [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] with self.rt.instance_claim(context, instance, node, allocs, [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 862.360221] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] return f(*args, **kwargs) [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] self._update(elevated, cn) [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] self._update_to_placement(context, compute_node, startup) [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] return attempt.get(self._wrap_exception) [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] six.reraise(self.value[0], self.value[1], self.value[2]) [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] raise value [ 862.360541] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] self.reportclient.update_from_provider_tree( [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] with catch_all(pd.uuid): [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] self.gen.throw(typ, value, traceback) [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] raise exception.ResourceProviderSyncFailed() [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 862.360963] env[69171]: ERROR nova.compute.manager [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] [ 862.361340] env[69171]: DEBUG nova.compute.utils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 862.363916] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.834s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.365462] env[69171]: INFO nova.compute.claims [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 862.369344] env[69171]: DEBUG nova.compute.manager [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Build of instance 6a0befb0-adf1-4907-845b-7e29f28f4f6a was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 862.370012] env[69171]: DEBUG nova.compute.manager [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 862.370342] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Acquiring lock "refresh_cache-6a0befb0-adf1-4907-845b-7e29f28f4f6a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.370574] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Acquired lock "refresh_cache-6a0befb0-adf1-4907-845b-7e29f28f4f6a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.370800] env[69171]: DEBUG nova.network.neutron [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 862.642713] env[69171]: INFO nova.compute.manager [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] [instance: 6071a99a-ff47-4161-8a1e-65385318891d] Took 1.03 seconds to deallocate network for instance. [ 862.893452] env[69171]: DEBUG nova.network.neutron [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 862.974155] env[69171]: DEBUG nova.network.neutron [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.397052] env[69171]: DEBUG nova.scheduler.client.report [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 863.409803] env[69171]: DEBUG nova.scheduler.client.report [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 863.410046] env[69171]: DEBUG nova.compute.provider_tree [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 863.420922] env[69171]: DEBUG nova.scheduler.client.report [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 863.439124] env[69171]: DEBUG nova.scheduler.client.report [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 863.478586] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Releasing lock "refresh_cache-6a0befb0-adf1-4907-845b-7e29f28f4f6a" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.478853] env[69171]: DEBUG nova.compute.manager [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 863.479076] env[69171]: DEBUG nova.compute.manager [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 863.479215] env[69171]: DEBUG nova.network.neutron [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 863.495538] env[69171]: DEBUG nova.network.neutron [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 863.603778] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce97e124-c068-4118-8080-1d9e24c96ac8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.611552] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34236ff-c343-4d79-9904-c73e9be3d1d1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.641590] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f79382af-37de-49b0-9be5-5c966bf51571 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.649449] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3457cc9a-af21-49bf-98f7-a72a8ab46fab {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.667552] env[69171]: DEBUG nova.compute.provider_tree [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 863.671196] env[69171]: INFO nova.scheduler.client.report [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Deleted allocations for instance 6071a99a-ff47-4161-8a1e-65385318891d [ 863.997779] env[69171]: DEBUG nova.network.neutron [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.178435] env[69171]: DEBUG oslo_concurrency.lockutils [None req-492ddeae-5dfd-448e-b651-76ec4dfc755f tempest-ServerPasswordTestJSON-32626778 tempest-ServerPasswordTestJSON-32626778-project-member] Lock "6071a99a-ff47-4161-8a1e-65385318891d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.585s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.187400] env[69171]: ERROR nova.scheduler.client.report [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [req-147a43de-a9db-4150-a437-49e2175f37a7] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-147a43de-a9db-4150-a437-49e2175f37a7"}]} [ 864.187620] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.824s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.188203] env[69171]: ERROR nova.compute.manager [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 864.188203] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Traceback (most recent call last): [ 864.188203] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 864.188203] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] yield [ 864.188203] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 864.188203] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] self.set_inventory_for_provider( [ 864.188203] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 864.188203] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-147a43de-a9db-4150-a437-49e2175f37a7"}]} [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] During handling of the above exception, another exception occurred: [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Traceback (most recent call last): [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] with self.rt.instance_claim(context, instance, node, allocs, [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 864.188496] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] return f(*args, **kwargs) [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] self._update(elevated, cn) [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] self._update_to_placement(context, compute_node, startup) [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] return attempt.get(self._wrap_exception) [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] six.reraise(self.value[0], self.value[1], self.value[2]) [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] raise value [ 864.188872] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] self.reportclient.update_from_provider_tree( [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] with catch_all(pd.uuid): [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] self.gen.throw(typ, value, traceback) [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] raise exception.ResourceProviderSyncFailed() [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 864.189323] env[69171]: ERROR nova.compute.manager [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] [ 864.189671] env[69171]: DEBUG nova.compute.utils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 864.190271] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.651s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.191736] env[69171]: INFO nova.compute.claims [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 864.194285] env[69171]: DEBUG nova.compute.manager [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Build of instance 1875b4bd-9024-40c0-9c43-244224904ce9 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 864.194680] env[69171]: DEBUG nova.compute.manager [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 864.194900] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Acquiring lock "refresh_cache-1875b4bd-9024-40c0-9c43-244224904ce9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.195053] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Acquired lock "refresh_cache-1875b4bd-9024-40c0-9c43-244224904ce9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.195220] env[69171]: DEBUG nova.network.neutron [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 864.500955] env[69171]: INFO nova.compute.manager [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] [instance: 6a0befb0-adf1-4907-845b-7e29f28f4f6a] Took 1.02 seconds to deallocate network for instance. [ 864.714969] env[69171]: DEBUG nova.network.neutron [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 864.802437] env[69171]: DEBUG nova.network.neutron [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.222276] env[69171]: DEBUG nova.scheduler.client.report [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 865.238590] env[69171]: DEBUG nova.scheduler.client.report [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 865.238882] env[69171]: DEBUG nova.compute.provider_tree [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 865.255541] env[69171]: DEBUG nova.scheduler.client.report [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 865.278101] env[69171]: DEBUG nova.scheduler.client.report [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 865.305082] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Releasing lock "refresh_cache-1875b4bd-9024-40c0-9c43-244224904ce9" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.305349] env[69171]: DEBUG nova.compute.manager [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 865.305536] env[69171]: DEBUG nova.compute.manager [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 865.305717] env[69171]: DEBUG nova.network.neutron [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 865.325016] env[69171]: DEBUG nova.network.neutron [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 865.480389] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d81cc51-92ca-4b1e-b9df-e122676d4042 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.488685] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d877d7-9d31-4b9c-b218-d6cc5f448db7 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.527701] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f9fb59-68c9-4644-8f8c-35f3b7f92c02 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.536387] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ad9791-858c-4403-9cf4-f554df96a1bc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.552014] env[69171]: DEBUG nova.compute.provider_tree [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 865.556521] env[69171]: INFO nova.scheduler.client.report [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Deleted allocations for instance 6a0befb0-adf1-4907-845b-7e29f28f4f6a [ 865.831681] env[69171]: DEBUG nova.network.neutron [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.065812] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0aae1df5-c1a9-4621-9839-e890e143a45d tempest-ServersNegativeTestJSON-1711513759 tempest-ServersNegativeTestJSON-1711513759-project-member] Lock "6a0befb0-adf1-4907-845b-7e29f28f4f6a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 39.655s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.089028] env[69171]: ERROR nova.scheduler.client.report [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [req-aca07921-427e-4050-b868-7f60a1b6b719] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-aca07921-427e-4050-b868-7f60a1b6b719"}]} [ 866.089028] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.897s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.089212] env[69171]: ERROR nova.compute.manager [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 866.089212] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Traceback (most recent call last): [ 866.089212] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 866.089212] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] yield [ 866.089212] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 866.089212] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] self.set_inventory_for_provider( [ 866.089212] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 866.089212] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-aca07921-427e-4050-b868-7f60a1b6b719"}]} [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] During handling of the above exception, another exception occurred: [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Traceback (most recent call last): [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] with self.rt.instance_claim(context, instance, node, allocs, [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 866.089458] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] return f(*args, **kwargs) [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] self._update(elevated, cn) [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] self._update_to_placement(context, compute_node, startup) [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] return attempt.get(self._wrap_exception) [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] raise value [ 866.089766] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] self.reportclient.update_from_provider_tree( [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] with catch_all(pd.uuid): [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] self.gen.throw(typ, value, traceback) [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] raise exception.ResourceProviderSyncFailed() [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 866.090217] env[69171]: ERROR nova.compute.manager [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] [ 866.090589] env[69171]: DEBUG nova.compute.utils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 866.090589] env[69171]: DEBUG nova.compute.manager [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Build of instance de9e4934-543e-433b-9094-ef2ce264a2f6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 866.090682] env[69171]: DEBUG nova.compute.manager [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 866.090964] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "refresh_cache-de9e4934-543e-433b-9094-ef2ce264a2f6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.091140] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquired lock "refresh_cache-de9e4934-543e-433b-9094-ef2ce264a2f6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.091299] env[69171]: DEBUG nova.network.neutron [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 866.092453] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.626s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.094071] env[69171]: INFO nova.compute.claims [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 866.336176] env[69171]: INFO nova.compute.manager [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] [instance: 1875b4bd-9024-40c0-9c43-244224904ce9] Took 1.03 seconds to deallocate network for instance. [ 866.625247] env[69171]: DEBUG nova.network.neutron [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 866.763606] env[69171]: DEBUG nova.network.neutron [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.136022] env[69171]: DEBUG nova.scheduler.client.report [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 867.154697] env[69171]: DEBUG nova.scheduler.client.report [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 867.154926] env[69171]: DEBUG nova.compute.provider_tree [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 867.174328] env[69171]: DEBUG nova.scheduler.client.report [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 867.199329] env[69171]: DEBUG nova.scheduler.client.report [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 867.273325] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Releasing lock "refresh_cache-de9e4934-543e-433b-9094-ef2ce264a2f6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.273555] env[69171]: DEBUG nova.compute.manager [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 867.273736] env[69171]: DEBUG nova.compute.manager [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 867.273972] env[69171]: DEBUG nova.network.neutron [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 867.313019] env[69171]: DEBUG nova.network.neutron [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 867.367145] env[69171]: INFO nova.scheduler.client.report [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Deleted allocations for instance 1875b4bd-9024-40c0-9c43-244224904ce9 [ 867.422242] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349f4dc2-cd5b-4b0a-b5d8-5335362adb09 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.430034] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a83c4cb-9f20-45a8-9baf-65d71834d5f5 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.465128] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b214d589-cf1a-4994-9da9-02b0019b7782 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.473672] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9a60b7c-6be7-403b-ae9f-74e2c992524d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.489909] env[69171]: DEBUG nova.compute.provider_tree [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 867.815258] env[69171]: DEBUG nova.network.neutron [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.878757] env[69171]: DEBUG oslo_concurrency.lockutils [None req-9b4f8bb3-bc2c-4832-b053-521dfdceac65 tempest-ServerRescueTestJSON-1722166625 tempest-ServerRescueTestJSON-1722166625-project-member] Lock "1875b4bd-9024-40c0-9c43-244224904ce9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.210s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.010922] env[69171]: ERROR nova.scheduler.client.report [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [req-68f64aa5-b1ba-4049-89a2-5fc1ab27e68f] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-68f64aa5-b1ba-4049-89a2-5fc1ab27e68f"}]} [ 868.011086] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.919s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.011744] env[69171]: ERROR nova.compute.manager [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 868.011744] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Traceback (most recent call last): [ 868.011744] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 868.011744] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] yield [ 868.011744] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 868.011744] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] self.set_inventory_for_provider( [ 868.011744] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 868.011744] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-68f64aa5-b1ba-4049-89a2-5fc1ab27e68f"}]} [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] During handling of the above exception, another exception occurred: [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Traceback (most recent call last): [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] with self.rt.instance_claim(context, instance, node, allocs, [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 868.012041] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] return f(*args, **kwargs) [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] self._update(elevated, cn) [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] self._update_to_placement(context, compute_node, startup) [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] return attempt.get(self._wrap_exception) [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] six.reraise(self.value[0], self.value[1], self.value[2]) [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] raise value [ 868.012455] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] self.reportclient.update_from_provider_tree( [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] with catch_all(pd.uuid): [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] self.gen.throw(typ, value, traceback) [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] raise exception.ResourceProviderSyncFailed() [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 868.012881] env[69171]: ERROR nova.compute.manager [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] [ 868.013303] env[69171]: DEBUG nova.compute.utils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 868.014469] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.665s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.016357] env[69171]: INFO nova.compute.claims [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 868.019855] env[69171]: DEBUG nova.compute.manager [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Build of instance a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 868.020346] env[69171]: DEBUG nova.compute.manager [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 868.020641] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "refresh_cache-a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.020832] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquired lock "refresh_cache-a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.021030] env[69171]: DEBUG nova.network.neutron [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 868.319331] env[69171]: INFO nova.compute.manager [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: de9e4934-543e-433b-9094-ef2ce264a2f6] Took 1.04 seconds to deallocate network for instance. [ 868.540824] env[69171]: DEBUG nova.network.neutron [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 868.676163] env[69171]: DEBUG nova.network.neutron [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.048041] env[69171]: DEBUG nova.scheduler.client.report [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 869.067845] env[69171]: DEBUG nova.scheduler.client.report [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 869.068175] env[69171]: DEBUG nova.compute.provider_tree [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.087869] env[69171]: DEBUG nova.scheduler.client.report [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 869.108975] env[69171]: DEBUG nova.scheduler.client.report [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 869.178606] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Releasing lock "refresh_cache-a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.178859] env[69171]: DEBUG nova.compute.manager [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 869.179062] env[69171]: DEBUG nova.compute.manager [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 869.179241] env[69171]: DEBUG nova.network.neutron [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 869.208771] env[69171]: DEBUG nova.network.neutron [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 869.279941] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40386cec-223a-4323-aa2e-ff332cf7437c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.292557] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4ce69f-cbbf-4acb-8bea-59c81226aa16 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.346021] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1513dbb0-f12a-4232-b846-890d89dcc109 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.356399] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af74e86c-f3e6-4093-8e38-51d925e394e0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.364647] env[69171]: INFO nova.scheduler.client.report [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Deleted allocations for instance de9e4934-543e-433b-9094-ef2ce264a2f6 [ 869.386492] env[69171]: DEBUG nova.compute.provider_tree [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.711909] env[69171]: DEBUG nova.network.neutron [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 869.894145] env[69171]: DEBUG oslo_concurrency.lockutils [None req-423ca3c1-fa33-4b53-8ec9-3028722f7f22 tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "de9e4934-543e-433b-9094-ef2ce264a2f6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.579s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.910737] env[69171]: ERROR nova.scheduler.client.report [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [req-5ffab7b0-9967-4ae7-b32e-abe2d0b97a44] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5ffab7b0-9967-4ae7-b32e-abe2d0b97a44"}]} [ 869.911209] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.897s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 869.911881] env[69171]: ERROR nova.compute.manager [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 869.911881] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Traceback (most recent call last): [ 869.911881] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 869.911881] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] yield [ 869.911881] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 869.911881] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] self.set_inventory_for_provider( [ 869.911881] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 869.911881] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-5ffab7b0-9967-4ae7-b32e-abe2d0b97a44"}]} [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] During handling of the above exception, another exception occurred: [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Traceback (most recent call last): [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] with self.rt.instance_claim(context, instance, node, allocs, [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 869.912186] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] return f(*args, **kwargs) [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] self._update(elevated, cn) [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] self._update_to_placement(context, compute_node, startup) [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] return attempt.get(self._wrap_exception) [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] six.reraise(self.value[0], self.value[1], self.value[2]) [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] raise value [ 869.912565] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] self.reportclient.update_from_provider_tree( [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] with catch_all(pd.uuid): [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] self.gen.throw(typ, value, traceback) [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] raise exception.ResourceProviderSyncFailed() [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 869.913052] env[69171]: ERROR nova.compute.manager [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] [ 869.913516] env[69171]: DEBUG nova.compute.utils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 869.913988] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.549s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 869.915679] env[69171]: INFO nova.compute.claims [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 869.918506] env[69171]: DEBUG nova.compute.manager [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Build of instance 44295293-0089-49c8-bdda-97ba3331c7d0 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 869.919012] env[69171]: DEBUG nova.compute.manager [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 869.919251] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "refresh_cache-44295293-0089-49c8-bdda-97ba3331c7d0" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 869.919408] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquired lock "refresh_cache-44295293-0089-49c8-bdda-97ba3331c7d0" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 869.919627] env[69171]: DEBUG nova.network.neutron [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 870.214612] env[69171]: INFO nova.compute.manager [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b] Took 1.03 seconds to deallocate network for instance. [ 870.442045] env[69171]: DEBUG nova.network.neutron [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 870.519839] env[69171]: DEBUG nova.network.neutron [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 870.944710] env[69171]: DEBUG nova.scheduler.client.report [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 870.957458] env[69171]: DEBUG nova.scheduler.client.report [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 870.957721] env[69171]: DEBUG nova.compute.provider_tree [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 870.968388] env[69171]: DEBUG nova.scheduler.client.report [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 870.983626] env[69171]: DEBUG nova.scheduler.client.report [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 871.021612] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Releasing lock "refresh_cache-44295293-0089-49c8-bdda-97ba3331c7d0" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 871.021916] env[69171]: DEBUG nova.compute.manager [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 871.022160] env[69171]: DEBUG nova.compute.manager [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 871.022373] env[69171]: DEBUG nova.network.neutron [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 871.040067] env[69171]: DEBUG nova.network.neutron [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 871.108852] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c98cc547-d9b5-4281-bc6d-c5c016988e23 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.116449] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7ef0f2-6549-4fdb-8f9c-0ab355ac5ebc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.145828] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8e3006f-425c-4c45-8572-88e271aa9c98 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.153632] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ba60a50-c233-4674-a685-f4b917300edf {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.168623] env[69171]: DEBUG nova.compute.provider_tree [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.243864] env[69171]: INFO nova.scheduler.client.report [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Deleted allocations for instance a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b [ 871.310441] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "6c3bcd59-b170-474f-abe2-d436dee0677d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.311141] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "6c3bcd59-b170-474f-abe2-d436dee0677d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.544630] env[69171]: DEBUG nova.network.neutron [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.691005] env[69171]: ERROR nova.scheduler.client.report [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [req-a16e4ba4-f6ec-42a0-9b27-a7e58f502df1] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a16e4ba4-f6ec-42a0-9b27-a7e58f502df1"}]} [ 871.691439] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.778s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.692195] env[69171]: ERROR nova.compute.manager [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 871.692195] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Traceback (most recent call last): [ 871.692195] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 871.692195] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] yield [ 871.692195] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 871.692195] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] self.set_inventory_for_provider( [ 871.692195] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 871.692195] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-a16e4ba4-f6ec-42a0-9b27-a7e58f502df1"}]} [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] During handling of the above exception, another exception occurred: [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Traceback (most recent call last): [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] with self.rt.instance_claim(context, instance, node, allocs, [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 871.692509] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] return f(*args, **kwargs) [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] self._update(elevated, cn) [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] self._update_to_placement(context, compute_node, startup) [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] return attempt.get(self._wrap_exception) [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] raise value [ 871.693025] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] self.reportclient.update_from_provider_tree( [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] with catch_all(pd.uuid): [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] self.gen.throw(typ, value, traceback) [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] raise exception.ResourceProviderSyncFailed() [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 871.693431] env[69171]: ERROR nova.compute.manager [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] [ 871.693776] env[69171]: DEBUG nova.compute.utils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 871.694291] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.348s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.695879] env[69171]: INFO nova.compute.claims [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.698670] env[69171]: DEBUG nova.compute.manager [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Build of instance 5001d2de-7fdd-479c-9a10-8b3a3e5862b6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 871.699091] env[69171]: DEBUG nova.compute.manager [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 871.699321] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Acquiring lock "refresh_cache-5001d2de-7fdd-479c-9a10-8b3a3e5862b6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.699569] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Acquired lock "refresh_cache-5001d2de-7fdd-479c-9a10-8b3a3e5862b6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.699761] env[69171]: DEBUG nova.network.neutron [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 871.752875] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5802861e-eae4-4fec-bfeb-53da75721111 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "a27c42c7-312a-43e3-bb1a-4f7f8f14ea1b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 40.433s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.813459] env[69171]: DEBUG nova.compute.manager [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 872.047336] env[69171]: INFO nova.compute.manager [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 44295293-0089-49c8-bdda-97ba3331c7d0] Took 1.02 seconds to deallocate network for instance. [ 872.221075] env[69171]: DEBUG nova.network.neutron [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 872.306338] env[69171]: DEBUG nova.network.neutron [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.335181] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.724492] env[69171]: DEBUG nova.scheduler.client.report [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 872.739258] env[69171]: DEBUG nova.scheduler.client.report [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 872.739486] env[69171]: DEBUG nova.compute.provider_tree [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 872.750377] env[69171]: DEBUG nova.scheduler.client.report [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 872.768844] env[69171]: DEBUG nova.scheduler.client.report [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 872.809489] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Releasing lock "refresh_cache-5001d2de-7fdd-479c-9a10-8b3a3e5862b6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 872.809784] env[69171]: DEBUG nova.compute.manager [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 872.810009] env[69171]: DEBUG nova.compute.manager [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 872.810219] env[69171]: DEBUG nova.network.neutron [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 872.831787] env[69171]: DEBUG nova.network.neutron [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 872.893188] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc7c1787-1ab0-44d9-875b-d5f9d880317a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.902260] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5708f516-a14d-4f19-b056-1af661a35b43 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.933111] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97f613a0-4003-4df5-887e-735716d2eaf4 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.941080] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac7468b9-4217-433e-82d4-78333d1231db {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.954340] env[69171]: DEBUG nova.compute.provider_tree [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 873.076613] env[69171]: INFO nova.scheduler.client.report [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Deleted allocations for instance 44295293-0089-49c8-bdda-97ba3331c7d0 [ 873.335013] env[69171]: DEBUG nova.network.neutron [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.476625] env[69171]: ERROR nova.scheduler.client.report [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [req-041a6c41-a509-4aa2-826a-6aa6a1f86ffc] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-041a6c41-a509-4aa2-826a-6aa6a1f86ffc"}]} [ 873.476990] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.783s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.477580] env[69171]: ERROR nova.compute.manager [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 873.477580] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Traceback (most recent call last): [ 873.477580] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 873.477580] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] yield [ 873.477580] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 873.477580] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] self.set_inventory_for_provider( [ 873.477580] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 873.477580] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-041a6c41-a509-4aa2-826a-6aa6a1f86ffc"}]} [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] During handling of the above exception, another exception occurred: [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Traceback (most recent call last): [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] with self.rt.instance_claim(context, instance, node, allocs, [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 873.477846] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] return f(*args, **kwargs) [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] self._update(elevated, cn) [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] self._update_to_placement(context, compute_node, startup) [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] return attempt.get(self._wrap_exception) [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] six.reraise(self.value[0], self.value[1], self.value[2]) [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] raise value [ 873.478260] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] self.reportclient.update_from_provider_tree( [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] with catch_all(pd.uuid): [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] self.gen.throw(typ, value, traceback) [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] raise exception.ResourceProviderSyncFailed() [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 873.478696] env[69171]: ERROR nova.compute.manager [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] [ 873.479096] env[69171]: DEBUG nova.compute.utils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 873.479967] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 13.337s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.480183] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.480371] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=69171) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 873.480691] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.150s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.482587] env[69171]: INFO nova.compute.claims [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 873.485842] env[69171]: DEBUG nova.compute.manager [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Build of instance 3cc37f29-ae73-4a6a-8bb1-88484c59d83c was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 873.486269] env[69171]: DEBUG nova.compute.manager [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 873.486498] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquiring lock "refresh_cache-3cc37f29-ae73-4a6a-8bb1-88484c59d83c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.486646] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Acquired lock "refresh_cache-3cc37f29-ae73-4a6a-8bb1-88484c59d83c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.486806] env[69171]: DEBUG nova.network.neutron [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 873.490029] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19064588-9627-431f-bc3f-0c97dc1eb4ec {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.499070] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6afa6a-93ef-4c04-983d-b81269b47076 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.515858] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12dc5508-6734-4095-b62e-7b4ed8f90032 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.523605] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70989d26-6b25-4b6e-8071-a598e3c3f280 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.554730] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180328MB free_disk=0GB free_vcpus=48 pci_devices=None {{(pid=69171) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 873.554875] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.584704] env[69171]: DEBUG oslo_concurrency.lockutils [None req-930eb941-306a-45b5-9793-9f9134f1d94b tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "44295293-0089-49c8-bdda-97ba3331c7d0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 37.058s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.837939] env[69171]: INFO nova.compute.manager [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] [instance: 5001d2de-7fdd-479c-9a10-8b3a3e5862b6] Took 1.03 seconds to deallocate network for instance. [ 874.005178] env[69171]: DEBUG nova.network.neutron [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.078313] env[69171]: DEBUG nova.network.neutron [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.403354] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "d7ec6342-5ebd-4333-9a7f-57da85f61cb4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.403667] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "d7ec6342-5ebd-4333-9a7f-57da85f61cb4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.510503] env[69171]: DEBUG nova.scheduler.client.report [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 874.525479] env[69171]: DEBUG nova.scheduler.client.report [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 874.525702] env[69171]: DEBUG nova.compute.provider_tree [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.538022] env[69171]: DEBUG nova.scheduler.client.report [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 874.557236] env[69171]: DEBUG nova.scheduler.client.report [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 874.580659] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Releasing lock "refresh_cache-3cc37f29-ae73-4a6a-8bb1-88484c59d83c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.580889] env[69171]: DEBUG nova.compute.manager [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 874.581098] env[69171]: DEBUG nova.compute.manager [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 874.581283] env[69171]: DEBUG nova.network.neutron [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 874.601023] env[69171]: DEBUG nova.network.neutron [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 874.684990] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57dad760-aa84-414a-86f8-007509e72a8e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.694426] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e12ee0c5-c77d-4220-b59f-4642658ff21c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.724247] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237de11c-21b4-4a98-aef5-e861bda344aa {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.731959] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1539048-0710-428b-b347-9ddec892ee8c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.745442] env[69171]: DEBUG nova.compute.provider_tree [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 874.865507] env[69171]: INFO nova.scheduler.client.report [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Deleted allocations for instance 5001d2de-7fdd-479c-9a10-8b3a3e5862b6 [ 874.905676] env[69171]: DEBUG nova.compute.manager [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 875.103219] env[69171]: DEBUG nova.network.neutron [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.167422] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "1d24d228-9cc0-4c93-a64b-5545d02575be" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.167657] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "1d24d228-9cc0-4c93-a64b-5545d02575be" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.267352] env[69171]: ERROR nova.scheduler.client.report [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [req-ac18751d-d12b-4d3e-888b-f93a0e0f39f9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ac18751d-d12b-4d3e-888b-f93a0e0f39f9"}]} [ 875.267700] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.787s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.268299] env[69171]: ERROR nova.compute.manager [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 875.268299] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Traceback (most recent call last): [ 875.268299] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 875.268299] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] yield [ 875.268299] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 875.268299] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] self.set_inventory_for_provider( [ 875.268299] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 875.268299] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-ac18751d-d12b-4d3e-888b-f93a0e0f39f9"}]} [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] During handling of the above exception, another exception occurred: [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Traceback (most recent call last): [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] with self.rt.instance_claim(context, instance, node, allocs, [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 875.268585] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] return f(*args, **kwargs) [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] self._update(elevated, cn) [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] self._update_to_placement(context, compute_node, startup) [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] return attempt.get(self._wrap_exception) [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] six.reraise(self.value[0], self.value[1], self.value[2]) [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] raise value [ 875.268971] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] self.reportclient.update_from_provider_tree( [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] with catch_all(pd.uuid): [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] self.gen.throw(typ, value, traceback) [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] raise exception.ResourceProviderSyncFailed() [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 875.269439] env[69171]: ERROR nova.compute.manager [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] [ 875.269847] env[69171]: DEBUG nova.compute.utils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 875.270283] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.033s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.271851] env[69171]: INFO nova.compute.claims [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.274482] env[69171]: DEBUG nova.compute.manager [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Build of instance a8c74dc9-6fb9-46f1-b558-2797ccfdfd16 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 875.274873] env[69171]: DEBUG nova.compute.manager [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 875.275120] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "refresh_cache-a8c74dc9-6fb9-46f1-b558-2797ccfdfd16" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.275250] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquired lock "refresh_cache-a8c74dc9-6fb9-46f1-b558-2797ccfdfd16" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.275406] env[69171]: DEBUG nova.network.neutron [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 875.372764] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01ae381d-d094-42ad-b7de-c11111ad1207 tempest-ServerRescueTestJSONUnderV235-2051269686 tempest-ServerRescueTestJSONUnderV235-2051269686-project-member] Lock "5001d2de-7fdd-479c-9a10-8b3a3e5862b6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 29.684s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.428080] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.607087] env[69171]: INFO nova.compute.manager [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] [instance: 3cc37f29-ae73-4a6a-8bb1-88484c59d83c] Took 1.03 seconds to deallocate network for instance. [ 875.670591] env[69171]: DEBUG nova.compute.manager [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 875.803937] env[69171]: DEBUG nova.network.neutron [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 875.939752] env[69171]: DEBUG nova.network.neutron [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.189429] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.304778] env[69171]: DEBUG nova.scheduler.client.report [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 876.319791] env[69171]: DEBUG nova.scheduler.client.report [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 876.320094] env[69171]: DEBUG nova.compute.provider_tree [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 876.332309] env[69171]: DEBUG nova.scheduler.client.report [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 876.352708] env[69171]: DEBUG nova.scheduler.client.report [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 876.442722] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Releasing lock "refresh_cache-a8c74dc9-6fb9-46f1-b558-2797ccfdfd16" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.443016] env[69171]: DEBUG nova.compute.manager [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 876.443184] env[69171]: DEBUG nova.compute.manager [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 876.443354] env[69171]: DEBUG nova.network.neutron [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 876.459736] env[69171]: DEBUG nova.network.neutron [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 876.484984] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cb5b09-e372-4275-9e3d-377199900039 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.493170] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabe3754-4ea8-4627-b9f0-375dc3235cb0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.526076] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4f63e5-e6be-4ec2-a6d3-ca3e82cca5fe {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.534592] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d92e87-f0c1-4420-a321-653676f1710f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.550068] env[69171]: DEBUG nova.compute.provider_tree [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 876.642279] env[69171]: INFO nova.scheduler.client.report [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Deleted allocations for instance 3cc37f29-ae73-4a6a-8bb1-88484c59d83c [ 876.964744] env[69171]: DEBUG nova.network.neutron [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.076769] env[69171]: ERROR nova.scheduler.client.report [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [req-225411d7-0042-4122-a9c9-44aef7988802] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-225411d7-0042-4122-a9c9-44aef7988802"}]} [ 877.077222] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.807s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.077879] env[69171]: ERROR nova.compute.manager [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 877.077879] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Traceback (most recent call last): [ 877.077879] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 877.077879] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] yield [ 877.077879] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 877.077879] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] self.set_inventory_for_provider( [ 877.077879] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 877.077879] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-225411d7-0042-4122-a9c9-44aef7988802"}]} [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] During handling of the above exception, another exception occurred: [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Traceback (most recent call last): [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] with self.rt.instance_claim(context, instance, node, allocs, [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 877.078235] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] return f(*args, **kwargs) [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] self._update(elevated, cn) [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] self._update_to_placement(context, compute_node, startup) [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] return attempt.get(self._wrap_exception) [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] six.reraise(self.value[0], self.value[1], self.value[2]) [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] raise value [ 877.078667] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] self.reportclient.update_from_provider_tree( [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] with catch_all(pd.uuid): [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] self.gen.throw(typ, value, traceback) [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] raise exception.ResourceProviderSyncFailed() [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 877.079099] env[69171]: ERROR nova.compute.manager [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] [ 877.082733] env[69171]: DEBUG nova.compute.utils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 877.084089] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.749s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.085569] env[69171]: INFO nova.compute.claims [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.088650] env[69171]: DEBUG nova.compute.manager [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Build of instance 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 877.089139] env[69171]: DEBUG nova.compute.manager [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 877.089400] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "refresh_cache-87abf9b1-06f4-412f-bb1d-f1e4cb77ef82" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.089548] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquired lock "refresh_cache-87abf9b1-06f4-412f-bb1d-f1e4cb77ef82" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.089718] env[69171]: DEBUG nova.network.neutron [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 877.153483] env[69171]: DEBUG oslo_concurrency.lockutils [None req-8f2a1c6a-53f5-4599-8d04-eab5c76e4e41 tempest-AttachVolumeTestJSON-591292491 tempest-AttachVolumeTestJSON-591292491-project-member] Lock "3cc37f29-ae73-4a6a-8bb1-88484c59d83c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.271s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.468929] env[69171]: INFO nova.compute.manager [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: a8c74dc9-6fb9-46f1-b558-2797ccfdfd16] Took 1.03 seconds to deallocate network for instance. [ 877.612434] env[69171]: DEBUG nova.network.neutron [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 877.704138] env[69171]: DEBUG nova.network.neutron [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.117593] env[69171]: DEBUG nova.scheduler.client.report [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 878.135336] env[69171]: DEBUG nova.scheduler.client.report [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 878.135336] env[69171]: DEBUG nova.compute.provider_tree [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 878.148954] env[69171]: DEBUG nova.scheduler.client.report [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 878.167714] env[69171]: DEBUG nova.scheduler.client.report [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 878.206764] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Releasing lock "refresh_cache-87abf9b1-06f4-412f-bb1d-f1e4cb77ef82" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.206956] env[69171]: DEBUG nova.compute.manager [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 878.207175] env[69171]: DEBUG nova.compute.manager [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 878.207349] env[69171]: DEBUG nova.network.neutron [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 878.232303] env[69171]: DEBUG nova.network.neutron [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 878.309568] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819e48cd-f146-428a-835c-45621179f2cb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.318203] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85ec81c2-a619-45d5-8593-4c16f06a8649 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.352023] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877a2d98-64ca-42cf-a3f0-956fbc30f72c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.360467] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d72f3c4-1685-408f-862b-27f223815e2b {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.374826] env[69171]: DEBUG nova.compute.provider_tree [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 878.507466] env[69171]: INFO nova.scheduler.client.report [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Deleted allocations for instance a8c74dc9-6fb9-46f1-b558-2797ccfdfd16 [ 878.736107] env[69171]: DEBUG nova.network.neutron [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.896064] env[69171]: ERROR nova.scheduler.client.report [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [req-f2b45bee-e5f9-4f53-bd03-8355048d52f6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f2b45bee-e5f9-4f53-bd03-8355048d52f6"}]} [ 878.896329] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.812s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 878.896920] env[69171]: ERROR nova.compute.manager [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 878.896920] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Traceback (most recent call last): [ 878.896920] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 878.896920] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] yield [ 878.896920] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 878.896920] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] self.set_inventory_for_provider( [ 878.896920] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 878.896920] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f2b45bee-e5f9-4f53-bd03-8355048d52f6"}]} [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] During handling of the above exception, another exception occurred: [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Traceback (most recent call last): [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] with self.rt.instance_claim(context, instance, node, allocs, [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 878.897181] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] return f(*args, **kwargs) [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] self._update(elevated, cn) [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] self._update_to_placement(context, compute_node, startup) [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] return attempt.get(self._wrap_exception) [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] six.reraise(self.value[0], self.value[1], self.value[2]) [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] raise value [ 878.897544] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] self.reportclient.update_from_provider_tree( [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] with catch_all(pd.uuid): [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] self.gen.throw(typ, value, traceback) [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] raise exception.ResourceProviderSyncFailed() [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 878.897949] env[69171]: ERROR nova.compute.manager [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] [ 878.898289] env[69171]: DEBUG nova.compute.utils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 878.898903] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.344s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 878.900564] env[69171]: DEBUG nova.compute.manager [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Build of instance 6c3bcd59-b170-474f-abe2-d436dee0677d was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 878.900961] env[69171]: DEBUG nova.compute.manager [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 878.901338] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "refresh_cache-6c3bcd59-b170-474f-abe2-d436dee0677d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.901511] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquired lock "refresh_cache-6c3bcd59-b170-474f-abe2-d436dee0677d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.901701] env[69171]: DEBUG nova.network.neutron [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 879.022851] env[69171]: DEBUG oslo_concurrency.lockutils [None req-f46fe076-88a9-458c-8a01-cb3ee578abf9 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "a8c74dc9-6fb9-46f1-b558-2797ccfdfd16" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.372s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.241832] env[69171]: INFO nova.compute.manager [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82] Took 1.03 seconds to deallocate network for instance. [ 879.576198] env[69171]: DEBUG nova.network.neutron [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 879.648945] env[69171]: DEBUG nova.network.neutron [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.933274] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f5a71cca-1e32-4c3c-9bb8-83e634a864f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.933506] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 8e6a9d73-6705-484c-82e0-70c6a6860211 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.933655] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 6a29d495-c938-41ad-97b6-1331b6eff589 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.933783] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.933903] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.934035] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 6c3bcd59-b170-474f-abe2-d436dee0677d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 879.998833] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "f81e9340-ea33-49e8-8297-a9b926b20da7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.999076] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "f81e9340-ea33-49e8-8297-a9b926b20da7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.151429] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Releasing lock "refresh_cache-6c3bcd59-b170-474f-abe2-d436dee0677d" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.151705] env[69171]: DEBUG nova.compute.manager [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 880.151900] env[69171]: DEBUG nova.compute.manager [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 880.152085] env[69171]: DEBUG nova.network.neutron [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 880.168127] env[69171]: DEBUG nova.network.neutron [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 880.270907] env[69171]: INFO nova.scheduler.client.report [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Deleted allocations for instance 87abf9b1-06f4-412f-bb1d-f1e4cb77ef82 [ 880.437575] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance d7ec6342-5ebd-4333-9a7f-57da85f61cb4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 880.501078] env[69171]: DEBUG nova.compute.manager [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 880.670588] env[69171]: DEBUG nova.network.neutron [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.781515] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0916bb2d-0298-430b-a2b9-d5057b382199 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "87abf9b1-06f4-412f-bb1d-f1e4cb77ef82" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.037s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.940622] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 1d24d228-9cc0-4c93-a64b-5545d02575be has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 881.023956] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.174273] env[69171]: INFO nova.compute.manager [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 6c3bcd59-b170-474f-abe2-d436dee0677d] Took 1.02 seconds to deallocate network for instance. [ 881.444088] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f81e9340-ea33-49e8-8297-a9b926b20da7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 881.444088] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 881.444354] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 881.461867] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 881.477209] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 881.477401] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.489425] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 881.509032] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 881.619992] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e45d659-339c-44e4-a83a-90c63f8cb584 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.631025] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f04adc-5edc-4b3e-a393-a6f288c19d34 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.666776] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b44e39e9-1aff-488b-9610-3d991c0c75eb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.674904] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-088d71bc-dcf9-44ed-89d1-785e97a138b1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.694400] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 882.214322] env[69171]: INFO nova.scheduler.client.report [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Deleted allocations for instance 6c3bcd59-b170-474f-abe2-d436dee0677d [ 882.220579] env[69171]: ERROR nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [req-d34167a1-666b-45fb-bd23-18279764c089] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d34167a1-666b-45fb-bd23-18279764c089"}]} [ 882.220579] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.322s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.221112] env[69171]: ERROR nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Error updating resources for node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28.: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 882.221112] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 882.221112] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 882.221112] env[69171]: ERROR nova.compute.manager yield [ 882.221112] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 882.221112] env[69171]: ERROR nova.compute.manager self.set_inventory_for_provider( [ 882.221112] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 882.221112] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 882.221112] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-d34167a1-666b-45fb-bd23-18279764c089"}]} [ 882.221112] env[69171]: ERROR nova.compute.manager [ 882.221112] env[69171]: ERROR nova.compute.manager During handling of the above exception, another exception occurred: [ 882.221112] env[69171]: ERROR nova.compute.manager [ 882.221506] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 10584, in _update_available_resource_for_node [ 882.221506] env[69171]: ERROR nova.compute.manager self.rt.update_available_resource(context, nodename, [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 935, in update_available_resource [ 882.221506] env[69171]: ERROR nova.compute.manager self._update_available_resource(context, resources, startup=startup) [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 882.221506] env[69171]: ERROR nova.compute.manager return f(*args, **kwargs) [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1066, in _update_available_resource [ 882.221506] env[69171]: ERROR nova.compute.manager self._update(context, cn, startup=startup) [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 882.221506] env[69171]: ERROR nova.compute.manager self._update_to_placement(context, compute_node, startup) [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 882.221506] env[69171]: ERROR nova.compute.manager return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 882.221506] env[69171]: ERROR nova.compute.manager return attempt.get(self._wrap_exception) [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 882.221506] env[69171]: ERROR nova.compute.manager six.reraise(self.value[0], self.value[1], self.value[2]) [ 882.221506] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 882.222424] env[69171]: ERROR nova.compute.manager raise value [ 882.222424] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 882.222424] env[69171]: ERROR nova.compute.manager attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 882.222424] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 882.222424] env[69171]: ERROR nova.compute.manager self.reportclient.update_from_provider_tree( [ 882.222424] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 882.222424] env[69171]: ERROR nova.compute.manager with catch_all(pd.uuid): [ 882.222424] env[69171]: ERROR nova.compute.manager File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 882.222424] env[69171]: ERROR nova.compute.manager self.gen.throw(typ, value, traceback) [ 882.222424] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 882.222424] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderSyncFailed() [ 882.222424] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 882.222424] env[69171]: ERROR nova.compute.manager [ 882.222424] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.793s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.223090] env[69171]: INFO nova.compute.claims [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.729702] env[69171]: DEBUG oslo_concurrency.lockutils [None req-3dbb43a3-4d45-4cd6-9b7f-483f6280b9ce tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "6c3bcd59-b170-474f-abe2-d436dee0677d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.418s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.187394] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "663c51b3-c350-4899-8ee3-724ce5b0bcfd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.187632] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "663c51b3-c350-4899-8ee3-724ce5b0bcfd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.247434] env[69171]: DEBUG nova.scheduler.client.report [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 883.260875] env[69171]: DEBUG nova.scheduler.client.report [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 883.261123] env[69171]: DEBUG nova.compute.provider_tree [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 883.272539] env[69171]: DEBUG nova.scheduler.client.report [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 883.291563] env[69171]: DEBUG nova.scheduler.client.report [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 883.406390] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af15f040-3d58-482b-85b1-13f528db2281 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.414499] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-874f5188-2979-4f01-90ba-ae1d96cb8c1f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.445569] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716f15d6-60ee-4286-b0b7-017139de5578 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.454027] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec3d040f-72f0-43e4-859f-d1cc9d5e027f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.467937] env[69171]: DEBUG nova.compute.provider_tree [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 883.531861] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "58879570-a539-49dc-90f3-28da84b11af6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.532118] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "58879570-a539-49dc-90f3-28da84b11af6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.690491] env[69171]: DEBUG nova.compute.manager [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 883.989304] env[69171]: ERROR nova.scheduler.client.report [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [req-6eaea3ce-0c1c-4cf5-b389-3e0dca7fc044] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6eaea3ce-0c1c-4cf5-b389-3e0dca7fc044"}]} [ 883.989695] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.768s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.990381] env[69171]: ERROR nova.compute.manager [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 883.990381] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Traceback (most recent call last): [ 883.990381] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 883.990381] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] yield [ 883.990381] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 883.990381] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] self.set_inventory_for_provider( [ 883.990381] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 883.990381] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6eaea3ce-0c1c-4cf5-b389-3e0dca7fc044"}]} [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] During handling of the above exception, another exception occurred: [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Traceback (most recent call last): [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] with self.rt.instance_claim(context, instance, node, allocs, [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 883.990712] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] return f(*args, **kwargs) [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] self._update(elevated, cn) [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] self._update_to_placement(context, compute_node, startup) [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] return attempt.get(self._wrap_exception) [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] six.reraise(self.value[0], self.value[1], self.value[2]) [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] raise value [ 883.990990] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] self.reportclient.update_from_provider_tree( [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] with catch_all(pd.uuid): [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] self.gen.throw(typ, value, traceback) [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] raise exception.ResourceProviderSyncFailed() [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 883.991745] env[69171]: ERROR nova.compute.manager [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] [ 883.992297] env[69171]: DEBUG nova.compute.utils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 883.992406] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.803s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.993993] env[69171]: INFO nova.compute.claims [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 883.996824] env[69171]: DEBUG nova.compute.manager [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Build of instance d7ec6342-5ebd-4333-9a7f-57da85f61cb4 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 883.997324] env[69171]: DEBUG nova.compute.manager [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 883.997562] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquiring lock "refresh_cache-d7ec6342-5ebd-4333-9a7f-57da85f61cb4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.997744] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Acquired lock "refresh_cache-d7ec6342-5ebd-4333-9a7f-57da85f61cb4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.997912] env[69171]: DEBUG nova.network.neutron [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 884.034442] env[69171]: DEBUG nova.compute.manager [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 884.211167] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.518738] env[69171]: DEBUG nova.network.neutron [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 884.554522] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 884.600881] env[69171]: DEBUG nova.network.neutron [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.022277] env[69171]: DEBUG nova.scheduler.client.report [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 885.038129] env[69171]: DEBUG nova.scheduler.client.report [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 885.038361] env[69171]: DEBUG nova.compute.provider_tree [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 885.048546] env[69171]: DEBUG nova.scheduler.client.report [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 885.066038] env[69171]: DEBUG nova.scheduler.client.report [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 885.103204] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Releasing lock "refresh_cache-d7ec6342-5ebd-4333-9a7f-57da85f61cb4" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.103450] env[69171]: DEBUG nova.compute.manager [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 885.103636] env[69171]: DEBUG nova.compute.manager [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 885.103859] env[69171]: DEBUG nova.network.neutron [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 885.118954] env[69171]: DEBUG nova.network.neutron [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 885.174390] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c82f05d7-209e-4948-89ac-12044dfeb255 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.182194] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54956342-04a5-4b5b-983f-0ba625e1e826 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.211624] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c485ce5-2b43-4d8c-97b1-139dfec7273e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.219154] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0ed7a9-22bf-4b28-9e2e-3fa9a4755d0c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.232329] env[69171]: DEBUG nova.compute.provider_tree [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 885.623378] env[69171]: DEBUG nova.network.neutron [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.753746] env[69171]: ERROR nova.scheduler.client.report [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [req-baecac24-0fa0-41cc-9879-d2bb9df9db08] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-baecac24-0fa0-41cc-9879-d2bb9df9db08"}]} [ 885.754164] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.762s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.754740] env[69171]: ERROR nova.compute.manager [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 885.754740] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Traceback (most recent call last): [ 885.754740] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 885.754740] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] yield [ 885.754740] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 885.754740] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] self.set_inventory_for_provider( [ 885.754740] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 885.754740] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-baecac24-0fa0-41cc-9879-d2bb9df9db08"}]} [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] During handling of the above exception, another exception occurred: [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Traceback (most recent call last): [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] with self.rt.instance_claim(context, instance, node, allocs, [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 885.755061] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] return f(*args, **kwargs) [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] self._update(elevated, cn) [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] self._update_to_placement(context, compute_node, startup) [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] return attempt.get(self._wrap_exception) [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] six.reraise(self.value[0], self.value[1], self.value[2]) [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] raise value [ 885.755381] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] self.reportclient.update_from_provider_tree( [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] with catch_all(pd.uuid): [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] self.gen.throw(typ, value, traceback) [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] raise exception.ResourceProviderSyncFailed() [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 885.755859] env[69171]: ERROR nova.compute.manager [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] [ 885.756267] env[69171]: DEBUG nova.compute.utils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 885.756648] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.733s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.758398] env[69171]: INFO nova.compute.claims [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.761011] env[69171]: DEBUG nova.compute.manager [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Build of instance 1d24d228-9cc0-4c93-a64b-5545d02575be was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 885.761419] env[69171]: DEBUG nova.compute.manager [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 885.761625] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "refresh_cache-1d24d228-9cc0-4c93-a64b-5545d02575be" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.761846] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquired lock "refresh_cache-1d24d228-9cc0-4c93-a64b-5545d02575be" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.762016] env[69171]: DEBUG nova.network.neutron [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 886.126726] env[69171]: INFO nova.compute.manager [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] [instance: d7ec6342-5ebd-4333-9a7f-57da85f61cb4] Took 1.02 seconds to deallocate network for instance. [ 886.284863] env[69171]: DEBUG nova.network.neutron [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 886.362063] env[69171]: DEBUG nova.network.neutron [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.787023] env[69171]: DEBUG nova.scheduler.client.report [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 886.800372] env[69171]: DEBUG nova.scheduler.client.report [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 886.800591] env[69171]: DEBUG nova.compute.provider_tree [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 886.814034] env[69171]: DEBUG nova.scheduler.client.report [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 886.831849] env[69171]: DEBUG nova.scheduler.client.report [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 886.865599] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Releasing lock "refresh_cache-1d24d228-9cc0-4c93-a64b-5545d02575be" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.865837] env[69171]: DEBUG nova.compute.manager [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 886.866117] env[69171]: DEBUG nova.compute.manager [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 886.866369] env[69171]: DEBUG nova.network.neutron [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 886.882844] env[69171]: DEBUG nova.network.neutron [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 886.963865] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c357182-8eb7-4762-8ad9-cd291dda2325 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.971920] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be6a42a1-3f03-4861-8dce-169f4ae358a2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.002193] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93628812-76db-4654-9e3d-3d5956f449f2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.010379] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06dbb66d-6b74-4b48-95af-5def656c10c2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.024507] env[69171]: DEBUG nova.compute.provider_tree [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 887.154921] env[69171]: INFO nova.scheduler.client.report [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Deleted allocations for instance d7ec6342-5ebd-4333-9a7f-57da85f61cb4 [ 887.385452] env[69171]: DEBUG nova.network.neutron [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 887.554934] env[69171]: ERROR nova.scheduler.client.report [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [req-35be8fe9-51f0-42b2-a9dc-d0c52b0de7fa] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-35be8fe9-51f0-42b2-a9dc-d0c52b0de7fa"}]} [ 887.555367] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.799s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.556044] env[69171]: ERROR nova.compute.manager [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 887.556044] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Traceback (most recent call last): [ 887.556044] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 887.556044] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] yield [ 887.556044] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 887.556044] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] self.set_inventory_for_provider( [ 887.556044] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 887.556044] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-35be8fe9-51f0-42b2-a9dc-d0c52b0de7fa"}]} [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] During handling of the above exception, another exception occurred: [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Traceback (most recent call last): [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] with self.rt.instance_claim(context, instance, node, allocs, [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 887.556336] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] return f(*args, **kwargs) [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] self._update(elevated, cn) [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] self._update_to_placement(context, compute_node, startup) [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] return attempt.get(self._wrap_exception) [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] six.reraise(self.value[0], self.value[1], self.value[2]) [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] raise value [ 887.556673] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] self.reportclient.update_from_provider_tree( [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] with catch_all(pd.uuid): [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] self.gen.throw(typ, value, traceback) [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] raise exception.ResourceProviderSyncFailed() [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 887.557162] env[69171]: ERROR nova.compute.manager [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] [ 887.557536] env[69171]: DEBUG nova.compute.utils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 887.558234] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.347s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.559967] env[69171]: INFO nova.compute.claims [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.563387] env[69171]: DEBUG nova.compute.manager [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Build of instance f81e9340-ea33-49e8-8297-a9b926b20da7 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 887.563942] env[69171]: DEBUG nova.compute.manager [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 887.564145] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "refresh_cache-f81e9340-ea33-49e8-8297-a9b926b20da7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.564318] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquired lock "refresh_cache-f81e9340-ea33-49e8-8297-a9b926b20da7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.564493] env[69171]: DEBUG nova.network.neutron [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 887.668567] env[69171]: DEBUG oslo_concurrency.lockutils [None req-01516df8-299d-4f23-999d-d3e6f936a869 tempest-AttachInterfacesTestJSON-395418482 tempest-AttachInterfacesTestJSON-395418482-project-member] Lock "d7ec6342-5ebd-4333-9a7f-57da85f61cb4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.265s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.888820] env[69171]: INFO nova.compute.manager [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 1d24d228-9cc0-4c93-a64b-5545d02575be] Took 1.02 seconds to deallocate network for instance. [ 888.090363] env[69171]: DEBUG nova.network.neutron [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 888.222577] env[69171]: DEBUG nova.network.neutron [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.591084] env[69171]: DEBUG nova.scheduler.client.report [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 888.605666] env[69171]: DEBUG nova.scheduler.client.report [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 888.605898] env[69171]: DEBUG nova.compute.provider_tree [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 888.617160] env[69171]: DEBUG nova.scheduler.client.report [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 888.636565] env[69171]: DEBUG nova.scheduler.client.report [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 888.725317] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Releasing lock "refresh_cache-f81e9340-ea33-49e8-8297-a9b926b20da7" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.726844] env[69171]: DEBUG nova.compute.manager [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 888.727245] env[69171]: DEBUG nova.compute.manager [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 888.727553] env[69171]: DEBUG nova.network.neutron [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 888.747407] env[69171]: DEBUG nova.network.neutron [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 888.758105] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf921c9e-6ec9-40f3-bdb4-42d52a0a18a1 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.767317] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af6bee79-d7af-4eb7-acab-7bd5e3f5baeb {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.801079] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a48e621-f1e2-4c9d-9284-ab6cfd283736 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.810286] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bacecead-61dd-4d27-a36e-d455db26a544 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.824821] env[69171]: DEBUG nova.compute.provider_tree [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 888.916799] env[69171]: INFO nova.scheduler.client.report [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Deleted allocations for instance 1d24d228-9cc0-4c93-a64b-5545d02575be [ 889.250877] env[69171]: DEBUG nova.network.neutron [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.347420] env[69171]: ERROR nova.scheduler.client.report [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [req-c2e7d013-3fbd-4581-ade3-12f8a4aa3f56] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c2e7d013-3fbd-4581-ade3-12f8a4aa3f56"}]} [ 889.347836] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.790s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.348474] env[69171]: ERROR nova.compute.manager [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 889.348474] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Traceback (most recent call last): [ 889.348474] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 889.348474] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] yield [ 889.348474] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 889.348474] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] self.set_inventory_for_provider( [ 889.348474] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 889.348474] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-c2e7d013-3fbd-4581-ade3-12f8a4aa3f56"}]} [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] During handling of the above exception, another exception occurred: [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Traceback (most recent call last): [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] with self.rt.instance_claim(context, instance, node, allocs, [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 889.348761] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] return f(*args, **kwargs) [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] self._update(elevated, cn) [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] self._update_to_placement(context, compute_node, startup) [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] return attempt.get(self._wrap_exception) [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] six.reraise(self.value[0], self.value[1], self.value[2]) [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] raise value [ 889.349112] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] self.reportclient.update_from_provider_tree( [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] with catch_all(pd.uuid): [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] self.gen.throw(typ, value, traceback) [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] raise exception.ResourceProviderSyncFailed() [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 889.349583] env[69171]: ERROR nova.compute.manager [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] [ 889.349954] env[69171]: DEBUG nova.compute.utils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 889.350522] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.796s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.351947] env[69171]: INFO nova.compute.claims [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.354837] env[69171]: DEBUG nova.compute.manager [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Build of instance 663c51b3-c350-4899-8ee3-724ce5b0bcfd was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 889.355330] env[69171]: DEBUG nova.compute.manager [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 889.355621] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "refresh_cache-663c51b3-c350-4899-8ee3-724ce5b0bcfd" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.355792] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquired lock "refresh_cache-663c51b3-c350-4899-8ee3-724ce5b0bcfd" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.355961] env[69171]: DEBUG nova.network.neutron [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 889.424754] env[69171]: DEBUG oslo_concurrency.lockutils [None req-88455cca-77e3-4c64-b4c7-191ff3f3d598 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "1d24d228-9cc0-4c93-a64b-5545d02575be" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.257s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.754691] env[69171]: INFO nova.compute.manager [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: f81e9340-ea33-49e8-8297-a9b926b20da7] Took 1.03 seconds to deallocate network for instance. [ 889.876202] env[69171]: DEBUG nova.network.neutron [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 889.959570] env[69171]: DEBUG nova.network.neutron [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.381307] env[69171]: DEBUG nova.scheduler.client.report [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 890.396249] env[69171]: DEBUG nova.scheduler.client.report [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 890.396480] env[69171]: DEBUG nova.compute.provider_tree [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.410170] env[69171]: DEBUG nova.scheduler.client.report [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 890.432207] env[69171]: DEBUG nova.scheduler.client.report [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 890.462870] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Releasing lock "refresh_cache-663c51b3-c350-4899-8ee3-724ce5b0bcfd" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.463180] env[69171]: DEBUG nova.compute.manager [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 890.463404] env[69171]: DEBUG nova.compute.manager [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 890.463582] env[69171]: DEBUG nova.network.neutron [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 890.481350] env[69171]: DEBUG nova.network.neutron [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 890.541147] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7511803b-9696-422e-aa8f-7a56dd845302 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.548984] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e69d2b2-b291-4986-9e49-739e96d0faa6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.579333] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b220a47-f21c-4ec5-a977-0e06cdcfc2f2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.586562] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2eadd3e-c9a1-448c-9668-2d47ffc933e8 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.599777] env[69171]: DEBUG nova.compute.provider_tree [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.700942] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "5d0324fd-8c20-435a-97f1-5d4e69bc875c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.701157] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "5d0324fd-8c20-435a-97f1-5d4e69bc875c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.781932] env[69171]: INFO nova.scheduler.client.report [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Deleted allocations for instance f81e9340-ea33-49e8-8297-a9b926b20da7 [ 890.984573] env[69171]: DEBUG nova.network.neutron [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.120321] env[69171]: ERROR nova.scheduler.client.report [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [req-fb8f381d-036e-43f8-b6f8-23d6e2fed60e] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fb8f381d-036e-43f8-b6f8-23d6e2fed60e"}]} [ 891.120651] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.770s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.121249] env[69171]: ERROR nova.compute.manager [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 891.121249] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] Traceback (most recent call last): [ 891.121249] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 891.121249] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] yield [ 891.121249] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 891.121249] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] self.set_inventory_for_provider( [ 891.121249] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 891.121249] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-fb8f381d-036e-43f8-b6f8-23d6e2fed60e"}]} [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] During handling of the above exception, another exception occurred: [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] Traceback (most recent call last): [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] with self.rt.instance_claim(context, instance, node, allocs, [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 891.121521] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] return f(*args, **kwargs) [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] self._update(elevated, cn) [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] self._update_to_placement(context, compute_node, startup) [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] return attempt.get(self._wrap_exception) [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] six.reraise(self.value[0], self.value[1], self.value[2]) [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] raise value [ 891.121887] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] self.reportclient.update_from_provider_tree( [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] with catch_all(pd.uuid): [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] self.gen.throw(typ, value, traceback) [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] raise exception.ResourceProviderSyncFailed() [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 891.122408] env[69171]: ERROR nova.compute.manager [instance: 58879570-a539-49dc-90f3-28da84b11af6] [ 891.122808] env[69171]: DEBUG nova.compute.utils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 891.123825] env[69171]: DEBUG nova.compute.manager [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Build of instance 58879570-a539-49dc-90f3-28da84b11af6 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 891.124299] env[69171]: DEBUG nova.compute.manager [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 891.124530] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquiring lock "refresh_cache-58879570-a539-49dc-90f3-28da84b11af6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.124676] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Acquired lock "refresh_cache-58879570-a539-49dc-90f3-28da84b11af6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.124834] env[69171]: DEBUG nova.network.neutron [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 891.203251] env[69171]: DEBUG nova.compute.manager [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 891.289299] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ba5b0472-9abd-407b-a91c-7a951b956b8d tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "f81e9340-ea33-49e8-8297-a9b926b20da7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.290s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.487601] env[69171]: INFO nova.compute.manager [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: 663c51b3-c350-4899-8ee3-724ce5b0bcfd] Took 1.02 seconds to deallocate network for instance. [ 891.644126] env[69171]: DEBUG nova.network.neutron [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 891.723674] env[69171]: DEBUG nova.network.neutron [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.727217] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.727475] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.729000] env[69171]: INFO nova.compute.claims [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 892.227182] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Releasing lock "refresh_cache-58879570-a539-49dc-90f3-28da84b11af6" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.227407] env[69171]: DEBUG nova.compute.manager [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 892.227640] env[69171]: DEBUG nova.compute.manager [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 892.227789] env[69171]: DEBUG nova.network.neutron [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 892.243855] env[69171]: DEBUG nova.network.neutron [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 892.354234] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "4512c56e-c8b7-41f2-92c6-ea2404556cb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.354623] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "4512c56e-c8b7-41f2-92c6-ea2404556cb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.515487] env[69171]: INFO nova.scheduler.client.report [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Deleted allocations for instance 663c51b3-c350-4899-8ee3-724ce5b0bcfd [ 892.746070] env[69171]: DEBUG nova.network.neutron [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.752102] env[69171]: DEBUG nova.scheduler.client.report [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 892.764097] env[69171]: DEBUG nova.scheduler.client.report [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 892.764351] env[69171]: DEBUG nova.compute.provider_tree [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 892.775012] env[69171]: DEBUG nova.scheduler.client.report [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 892.792478] env[69171]: DEBUG nova.scheduler.client.report [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 892.856844] env[69171]: DEBUG nova.compute.manager [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 892.882820] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cb77737-1182-4c8f-a5c8-bd8db26c1b06 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.891045] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd2356c3-698d-4437-acfe-6816937e5c0e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.923607] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5854a02b-985c-4887-b902-46ebebeb4284 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.932169] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-caf60909-9164-4c0a-af49-f8dad30f9363 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.945989] env[69171]: DEBUG nova.compute.provider_tree [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 893.022959] env[69171]: DEBUG oslo_concurrency.lockutils [None req-ca3cb7ef-fc29-4d2c-b97f-e6783aaed994 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "663c51b3-c350-4899-8ee3-724ce5b0bcfd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 9.835s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.248949] env[69171]: INFO nova.compute.manager [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] [instance: 58879570-a539-49dc-90f3-28da84b11af6] Took 1.02 seconds to deallocate network for instance. [ 893.374301] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.467329] env[69171]: ERROR nova.scheduler.client.report [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [req-e1fb6ebe-7ffc-4372-8d06-521d081c8501] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e1fb6ebe-7ffc-4372-8d06-521d081c8501"}]} [ 893.467664] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.740s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.468247] env[69171]: ERROR nova.compute.manager [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 893.468247] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Traceback (most recent call last): [ 893.468247] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 893.468247] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] yield [ 893.468247] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 893.468247] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] self.set_inventory_for_provider( [ 893.468247] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 893.468247] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-e1fb6ebe-7ffc-4372-8d06-521d081c8501"}]} [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] During handling of the above exception, another exception occurred: [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Traceback (most recent call last): [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] with self.rt.instance_claim(context, instance, node, allocs, [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 893.468560] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] return f(*args, **kwargs) [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] self._update(elevated, cn) [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] self._update_to_placement(context, compute_node, startup) [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] return attempt.get(self._wrap_exception) [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] six.reraise(self.value[0], self.value[1], self.value[2]) [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] raise value [ 893.468919] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] self.reportclient.update_from_provider_tree( [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] with catch_all(pd.uuid): [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] self.gen.throw(typ, value, traceback) [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] raise exception.ResourceProviderSyncFailed() [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 893.469404] env[69171]: ERROR nova.compute.manager [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] [ 893.469804] env[69171]: DEBUG nova.compute.utils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 893.470291] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.096s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.472216] env[69171]: INFO nova.compute.claims [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.474662] env[69171]: DEBUG nova.compute.manager [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Build of instance 5d0324fd-8c20-435a-97f1-5d4e69bc875c was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 893.475072] env[69171]: DEBUG nova.compute.manager [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 893.475300] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "refresh_cache-5d0324fd-8c20-435a-97f1-5d4e69bc875c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.475448] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquired lock "refresh_cache-5d0324fd-8c20-435a-97f1-5d4e69bc875c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.475604] env[69171]: DEBUG nova.network.neutron [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 893.997507] env[69171]: DEBUG nova.network.neutron [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.102065] env[69171]: DEBUG nova.network.neutron [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.283376] env[69171]: INFO nova.scheduler.client.report [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Deleted allocations for instance 58879570-a539-49dc-90f3-28da84b11af6 [ 894.499423] env[69171]: DEBUG nova.scheduler.client.report [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 894.512969] env[69171]: DEBUG nova.scheduler.client.report [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 894.513218] env[69171]: DEBUG nova.compute.provider_tree [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 894.526047] env[69171]: DEBUG nova.scheduler.client.report [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 894.544358] env[69171]: DEBUG nova.scheduler.client.report [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 894.605435] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Releasing lock "refresh_cache-5d0324fd-8c20-435a-97f1-5d4e69bc875c" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 894.605783] env[69171]: DEBUG nova.compute.manager [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 894.605944] env[69171]: DEBUG nova.compute.manager [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 894.606131] env[69171]: DEBUG nova.network.neutron [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 894.621422] env[69171]: DEBUG nova.network.neutron [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 894.636651] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6297a5-2d63-4f67-bcdf-ae9d8712ce18 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.645302] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83af0921-7174-4119-859c-a6b010575869 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.676528] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29ca0d3f-f962-4859-877b-3a19140c3280 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.685149] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a44aef3c-29f4-4743-a5be-8fef02f86a55 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.698672] env[69171]: DEBUG nova.compute.provider_tree [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 894.792042] env[69171]: DEBUG oslo_concurrency.lockutils [None req-0f2f9aa8-7fc3-4e07-9ba8-b2ba7c93b79d tempest-ServersTestJSON-245932638 tempest-ServersTestJSON-245932638-project-member] Lock "58879570-a539-49dc-90f3-28da84b11af6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 11.260s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.126688] env[69171]: DEBUG nova.network.neutron [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.219831] env[69171]: ERROR nova.scheduler.client.report [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [req-1a68d481-37f5-460a-aab3-7c2f8abc699d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1a68d481-37f5-460a-aab3-7c2f8abc699d"}]} [ 895.220183] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.750s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.220764] env[69171]: ERROR nova.compute.manager [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 895.220764] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Traceback (most recent call last): [ 895.220764] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 895.220764] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] yield [ 895.220764] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 895.220764] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] self.set_inventory_for_provider( [ 895.220764] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 895.220764] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-1a68d481-37f5-460a-aab3-7c2f8abc699d"}]} [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] During handling of the above exception, another exception occurred: [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Traceback (most recent call last): [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] with self.rt.instance_claim(context, instance, node, allocs, [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 895.221044] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] return f(*args, **kwargs) [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] self._update(elevated, cn) [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] self._update_to_placement(context, compute_node, startup) [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] return attempt.get(self._wrap_exception) [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] six.reraise(self.value[0], self.value[1], self.value[2]) [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] raise value [ 895.221408] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] self.reportclient.update_from_provider_tree( [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] with catch_all(pd.uuid): [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] self.gen.throw(typ, value, traceback) [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] raise exception.ResourceProviderSyncFailed() [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 895.221892] env[69171]: ERROR nova.compute.manager [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] [ 895.222269] env[69171]: DEBUG nova.compute.utils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 895.223192] env[69171]: DEBUG nova.compute.manager [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Build of instance 4512c56e-c8b7-41f2-92c6-ea2404556cb2 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 895.223579] env[69171]: DEBUG nova.compute.manager [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 895.223802] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquiring lock "refresh_cache-4512c56e-c8b7-41f2-92c6-ea2404556cb2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.223949] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Acquired lock "refresh_cache-4512c56e-c8b7-41f2-92c6-ea2404556cb2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.224121] env[69171]: DEBUG nova.network.neutron [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 895.526841] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "d99e2135-c6fb-4ecb-b6f5-7bc2c5769826" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.527091] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "d99e2135-c6fb-4ecb-b6f5-7bc2c5769826" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.629577] env[69171]: INFO nova.compute.manager [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: 5d0324fd-8c20-435a-97f1-5d4e69bc875c] Took 1.02 seconds to deallocate network for instance. [ 895.742248] env[69171]: DEBUG nova.network.neutron [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 895.827019] env[69171]: DEBUG nova.network.neutron [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.029582] env[69171]: DEBUG nova.compute.manager [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 896.329928] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Releasing lock "refresh_cache-4512c56e-c8b7-41f2-92c6-ea2404556cb2" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.330213] env[69171]: DEBUG nova.compute.manager [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 896.330403] env[69171]: DEBUG nova.compute.manager [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 896.330568] env[69171]: DEBUG nova.network.neutron [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 896.345675] env[69171]: DEBUG nova.network.neutron [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 896.556546] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.556994] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.559424] env[69171]: INFO nova.compute.claims [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 896.662371] env[69171]: INFO nova.scheduler.client.report [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Deleted allocations for instance 5d0324fd-8c20-435a-97f1-5d4e69bc875c [ 896.848235] env[69171]: DEBUG nova.network.neutron [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.170366] env[69171]: DEBUG oslo_concurrency.lockutils [None req-e7d2a01e-36e0-4e1e-906c-7048e578bf07 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "5d0324fd-8c20-435a-97f1-5d4e69bc875c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 6.469s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.351726] env[69171]: INFO nova.compute.manager [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] [instance: 4512c56e-c8b7-41f2-92c6-ea2404556cb2] Took 1.02 seconds to deallocate network for instance. [ 897.587132] env[69171]: DEBUG nova.scheduler.client.report [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 897.601776] env[69171]: DEBUG nova.scheduler.client.report [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 897.601972] env[69171]: DEBUG nova.compute.provider_tree [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 897.614039] env[69171]: DEBUG nova.scheduler.client.report [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 897.636239] env[69171]: DEBUG nova.scheduler.client.report [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 897.730140] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d4f7f3e-2f33-4ac8-bc6e-f080dddbbd6e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.739058] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb223cc7-33d5-44ff-ae60-16e3448883ab {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.775653] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47598aec-fb33-4b78-a11e-a015539cb0b0 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.784461] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec7e37f7-4c79-464b-ae01-e7eb1a166cca {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.803159] env[69171]: DEBUG nova.compute.provider_tree [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 898.332152] env[69171]: ERROR nova.scheduler.client.report [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [req-2f481a8a-ffc3-4a73-8a51-10cf7d784c75] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2f481a8a-ffc3-4a73-8a51-10cf7d784c75"}]} [ 898.332152] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.773s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.332294] env[69171]: ERROR nova.compute.manager [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 898.332294] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Traceback (most recent call last): [ 898.332294] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 898.332294] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] yield [ 898.332294] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 898.332294] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] self.set_inventory_for_provider( [ 898.332294] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 898.332294] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-2f481a8a-ffc3-4a73-8a51-10cf7d784c75"}]} [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] During handling of the above exception, another exception occurred: [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Traceback (most recent call last): [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] with self.rt.instance_claim(context, instance, node, allocs, [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 898.332510] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] return f(*args, **kwargs) [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] self._update(elevated, cn) [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] self._update_to_placement(context, compute_node, startup) [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] return attempt.get(self._wrap_exception) [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] six.reraise(self.value[0], self.value[1], self.value[2]) [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] raise value [ 898.332827] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] self.reportclient.update_from_provider_tree( [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] with catch_all(pd.uuid): [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] self.gen.throw(typ, value, traceback) [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] raise exception.ResourceProviderSyncFailed() [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 898.333230] env[69171]: ERROR nova.compute.manager [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] [ 898.333693] env[69171]: DEBUG nova.compute.utils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 898.337023] env[69171]: DEBUG nova.compute.manager [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Build of instance d99e2135-c6fb-4ecb-b6f5-7bc2c5769826 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 898.337023] env[69171]: DEBUG nova.compute.manager [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 898.337023] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquiring lock "refresh_cache-d99e2135-c6fb-4ecb-b6f5-7bc2c5769826" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.337023] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Acquired lock "refresh_cache-d99e2135-c6fb-4ecb-b6f5-7bc2c5769826" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.337292] env[69171]: DEBUG nova.network.neutron [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 898.388122] env[69171]: INFO nova.scheduler.client.report [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Deleted allocations for instance 4512c56e-c8b7-41f2-92c6-ea2404556cb2 [ 898.859167] env[69171]: DEBUG nova.network.neutron [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 898.872237] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 898.872496] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.894037] env[69171]: DEBUG oslo_concurrency.lockutils [None req-5090ebde-9c6a-4570-9104-5cbe5044faa3 tempest-ServerDiskConfigTestJSON-494826628 tempest-ServerDiskConfigTestJSON-494826628-project-member] Lock "4512c56e-c8b7-41f2-92c6-ea2404556cb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 6.539s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.952830] env[69171]: DEBUG nova.network.neutron [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.378062] env[69171]: DEBUG nova.compute.manager [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 899.455284] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Releasing lock "refresh_cache-d99e2135-c6fb-4ecb-b6f5-7bc2c5769826" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.455588] env[69171]: DEBUG nova.compute.manager [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 899.455712] env[69171]: DEBUG nova.compute.manager [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 899.455899] env[69171]: DEBUG nova.network.neutron [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 899.476776] env[69171]: DEBUG nova.network.neutron [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 899.904036] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.904036] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.905386] env[69171]: INFO nova.compute.claims [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 899.979794] env[69171]: DEBUG nova.network.neutron [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 900.482587] env[69171]: INFO nova.compute.manager [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] [instance: d99e2135-c6fb-4ecb-b6f5-7bc2c5769826] Took 1.03 seconds to deallocate network for instance. [ 900.942896] env[69171]: DEBUG nova.scheduler.client.report [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 900.965467] env[69171]: DEBUG nova.scheduler.client.report [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 900.965700] env[69171]: DEBUG nova.compute.provider_tree [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 900.968091] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquiring lock "c1b91842-08db-403f-939e-d98fc9197aab" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.968347] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "c1b91842-08db-403f-939e-d98fc9197aab" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.982859] env[69171]: DEBUG nova.scheduler.client.report [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 901.005675] env[69171]: DEBUG nova.scheduler.client.report [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 901.109242] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef07f5fa-f68f-4efb-a43c-992e8c5078ac {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.118246] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8340f191-0940-4916-83d6-8fd5d29ec33e {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.148202] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d878a29f-0892-44d3-a66b-c445f563483c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.156474] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46da1483-efd3-45e2-97de-f37a9ad98b71 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.171029] env[69171]: DEBUG nova.compute.provider_tree [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 901.473099] env[69171]: DEBUG nova.compute.manager [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 901.517506] env[69171]: INFO nova.scheduler.client.report [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Deleted allocations for instance d99e2135-c6fb-4ecb-b6f5-7bc2c5769826 [ 901.692948] env[69171]: ERROR nova.scheduler.client.report [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [req-f49d2b2d-895d-4018-9f61-01e312d3456b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f49d2b2d-895d-4018-9f61-01e312d3456b"}]} [ 901.693346] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.789s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.693986] env[69171]: ERROR nova.compute.manager [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 901.693986] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Traceback (most recent call last): [ 901.693986] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 901.693986] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] yield [ 901.693986] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 901.693986] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] self.set_inventory_for_provider( [ 901.693986] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 901.693986] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-f49d2b2d-895d-4018-9f61-01e312d3456b"}]} [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] During handling of the above exception, another exception occurred: [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Traceback (most recent call last): [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] with self.rt.instance_claim(context, instance, node, allocs, [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 901.694288] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] return f(*args, **kwargs) [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] self._update(elevated, cn) [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] self._update_to_placement(context, compute_node, startup) [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] return attempt.get(self._wrap_exception) [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] six.reraise(self.value[0], self.value[1], self.value[2]) [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] raise value [ 901.694622] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] self.reportclient.update_from_provider_tree( [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] with catch_all(pd.uuid): [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] self.gen.throw(typ, value, traceback) [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] raise exception.ResourceProviderSyncFailed() [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 901.695134] env[69171]: ERROR nova.compute.manager [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] [ 901.695637] env[69171]: DEBUG nova.compute.utils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 901.696564] env[69171]: DEBUG nova.compute.manager [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Build of instance f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 901.696966] env[69171]: DEBUG nova.compute.manager [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 901.697377] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquiring lock "refresh_cache-f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.697377] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Acquired lock "refresh_cache-f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.697551] env[69171]: DEBUG nova.network.neutron [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 901.994369] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.994615] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.996247] env[69171]: INFO nova.compute.claims [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 902.025944] env[69171]: DEBUG oslo_concurrency.lockutils [None req-10e43b17-da7e-48d0-9b60-50a2bf6cff09 tempest-AttachVolumeNegativeTest-1809327485 tempest-AttachVolumeNegativeTest-1809327485-project-member] Lock "d99e2135-c6fb-4ecb-b6f5-7bc2c5769826" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 6.499s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.216996] env[69171]: DEBUG nova.network.neutron [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 902.307057] env[69171]: DEBUG nova.network.neutron [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.810679] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Releasing lock "refresh_cache-f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.810970] env[69171]: DEBUG nova.compute.manager [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 902.811139] env[69171]: DEBUG nova.compute.manager [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 902.811280] env[69171]: DEBUG nova.network.neutron [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 902.836032] env[69171]: DEBUG nova.network.neutron [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 903.026645] env[69171]: DEBUG nova.scheduler.client.report [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 903.048692] env[69171]: DEBUG nova.scheduler.client.report [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 903.048925] env[69171]: DEBUG nova.compute.provider_tree [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 903.063036] env[69171]: DEBUG nova.scheduler.client.report [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 903.085523] env[69171]: DEBUG nova.scheduler.client.report [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 903.172769] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b61fc6-95d2-4b8c-af96-86ae315b4e81 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.180980] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1526747f-10bd-4e3e-88be-10627e3a85fc {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.220368] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59470533-9eea-4042-b509-b78bca39dcf3 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.228560] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554a54f7-de6c-4a78-abb6-8256be3ce3da {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.243191] env[69171]: DEBUG nova.compute.provider_tree [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 903.338966] env[69171]: DEBUG nova.network.neutron [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.778264] env[69171]: ERROR nova.scheduler.client.report [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [req-57cb3daa-2a0a-464f-b9f5-bd2c66052cdd] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-57cb3daa-2a0a-464f-b9f5-bd2c66052cdd"}]} [ 903.778264] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.782s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.778468] env[69171]: ERROR nova.compute.manager [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 903.778468] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] Traceback (most recent call last): [ 903.778468] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 903.778468] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] yield [ 903.778468] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 903.778468] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] self.set_inventory_for_provider( [ 903.778468] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 903.778468] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-57cb3daa-2a0a-464f-b9f5-bd2c66052cdd"}]} [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] During handling of the above exception, another exception occurred: [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] Traceback (most recent call last): [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] with self.rt.instance_claim(context, instance, node, allocs, [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 903.778713] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] return f(*args, **kwargs) [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] self._update(elevated, cn) [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] self._update_to_placement(context, compute_node, startup) [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] return attempt.get(self._wrap_exception) [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] six.reraise(self.value[0], self.value[1], self.value[2]) [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] raise value [ 903.779026] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] self.reportclient.update_from_provider_tree( [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] with catch_all(pd.uuid): [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] self.gen.throw(typ, value, traceback) [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] raise exception.ResourceProviderSyncFailed() [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 903.779438] env[69171]: ERROR nova.compute.manager [instance: c1b91842-08db-403f-939e-d98fc9197aab] [ 903.779968] env[69171]: DEBUG nova.compute.utils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 903.783335] env[69171]: DEBUG nova.compute.manager [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Build of instance c1b91842-08db-403f-939e-d98fc9197aab was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 903.783335] env[69171]: DEBUG nova.compute.manager [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 903.783335] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquiring lock "refresh_cache-c1b91842-08db-403f-939e-d98fc9197aab" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.783335] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquired lock "refresh_cache-c1b91842-08db-403f-939e-d98fc9197aab" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.783528] env[69171]: DEBUG nova.network.neutron [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 903.842283] env[69171]: INFO nova.compute.manager [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] [instance: f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11] Took 1.03 seconds to deallocate network for instance. [ 904.304031] env[69171]: DEBUG nova.network.neutron [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 904.409878] env[69171]: DEBUG nova.network.neutron [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.881555] env[69171]: INFO nova.scheduler.client.report [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Deleted allocations for instance f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11 [ 904.913064] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Releasing lock "refresh_cache-c1b91842-08db-403f-939e-d98fc9197aab" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.913064] env[69171]: DEBUG nova.compute.manager [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 904.913354] env[69171]: DEBUG nova.compute.manager [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 904.915514] env[69171]: DEBUG nova.network.neutron [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 904.944077] env[69171]: DEBUG nova.network.neutron [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 905.393949] env[69171]: DEBUG oslo_concurrency.lockutils [None req-fcfc1f41-b030-4db4-a67f-a3e2897c1856 tempest-DeleteServersTestJSON-1769310840 tempest-DeleteServersTestJSON-1769310840-project-member] Lock "f6ad2d63-6fd6-4af6-8bd2-c0d5d4a5dd11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 6.521s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.446751] env[69171]: DEBUG nova.network.neutron [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.621496] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Acquiring lock "ef24399e-4d30-4415-b880-c0b0dccdc461" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.621776] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Lock "ef24399e-4d30-4415-b880-c0b0dccdc461" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.950293] env[69171]: INFO nova.compute.manager [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: c1b91842-08db-403f-939e-d98fc9197aab] Took 1.04 seconds to deallocate network for instance. [ 906.123819] env[69171]: DEBUG nova.compute.manager [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 906.648376] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.648619] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.650505] env[69171]: INFO nova.compute.claims [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.979940] env[69171]: INFO nova.scheduler.client.report [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Deleted allocations for instance c1b91842-08db-403f-939e-d98fc9197aab [ 907.487745] env[69171]: DEBUG oslo_concurrency.lockutils [None req-853e8e56-eded-4a45-b881-23527981ba63 tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "c1b91842-08db-403f-939e-d98fc9197aab" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 6.519s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.675867] env[69171]: DEBUG nova.scheduler.client.report [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 907.690024] env[69171]: DEBUG nova.scheduler.client.report [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 907.691103] env[69171]: DEBUG nova.compute.provider_tree [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 907.701673] env[69171]: DEBUG nova.scheduler.client.report [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 907.719874] env[69171]: DEBUG nova.scheduler.client.report [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 907.789578] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa6f27bc-ef8d-40fe-bf1d-ce6c98fdd1ee {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.799231] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3929119-0bd2-4ec5-9580-7e964c84aa8c {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.830146] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81825884-5004-47cf-8e3e-2a3ad8bee347 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.837903] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb8159d2-8d07-4c0f-b09f-7c0cd2272236 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.852407] env[69171]: DEBUG nova.compute.provider_tree [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 908.375035] env[69171]: ERROR nova.scheduler.client.report [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [req-de3b7724-1601-480b-a7b8-ae3aaede6bc4] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-de3b7724-1601-480b-a7b8-ae3aaede6bc4"}]} [ 908.375393] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.727s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.376353] env[69171]: ERROR nova.compute.manager [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 908.376353] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Traceback (most recent call last): [ 908.376353] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 908.376353] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] yield [ 908.376353] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 908.376353] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] self.set_inventory_for_provider( [ 908.376353] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 908.376353] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-de3b7724-1601-480b-a7b8-ae3aaede6bc4"}]} [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] During handling of the above exception, another exception occurred: [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Traceback (most recent call last): [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] with self.rt.instance_claim(context, instance, node, allocs, [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 908.376558] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] return f(*args, **kwargs) [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] self._update(elevated, cn) [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] self._update_to_placement(context, compute_node, startup) [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] return attempt.get(self._wrap_exception) [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] six.reraise(self.value[0], self.value[1], self.value[2]) [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] raise value [ 908.376779] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] self.reportclient.update_from_provider_tree( [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] with catch_all(pd.uuid): [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] self.gen.throw(typ, value, traceback) [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] raise exception.ResourceProviderSyncFailed() [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 908.377070] env[69171]: ERROR nova.compute.manager [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] [ 908.377346] env[69171]: DEBUG nova.compute.utils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 908.378700] env[69171]: DEBUG nova.compute.manager [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Build of instance ef24399e-4d30-4415-b880-c0b0dccdc461 was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 908.379106] env[69171]: DEBUG nova.compute.manager [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 908.379338] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Acquiring lock "refresh_cache-ef24399e-4d30-4415-b880-c0b0dccdc461" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 908.379484] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Acquired lock "refresh_cache-ef24399e-4d30-4415-b880-c0b0dccdc461" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 908.379644] env[69171]: DEBUG nova.network.neutron [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 908.902394] env[69171]: DEBUG nova.network.neutron [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 909.023136] env[69171]: DEBUG nova.network.neutron [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.526222] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Releasing lock "refresh_cache-ef24399e-4d30-4415-b880-c0b0dccdc461" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 909.526558] env[69171]: DEBUG nova.compute.manager [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 909.526659] env[69171]: DEBUG nova.compute.manager [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 909.526831] env[69171]: DEBUG nova.network.neutron [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 909.545063] env[69171]: DEBUG nova.network.neutron [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 910.048079] env[69171]: DEBUG nova.network.neutron [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.437735] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquiring lock "1214f461-59f9-4a21-ae58-ee4803622e8e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.437969] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "1214f461-59f9-4a21-ae58-ee4803622e8e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.551135] env[69171]: INFO nova.compute.manager [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] [instance: ef24399e-4d30-4415-b880-c0b0dccdc461] Took 1.02 seconds to deallocate network for instance. [ 910.939910] env[69171]: DEBUG nova.compute.manager [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Starting instance... {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2433}} [ 911.471615] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.472015] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.473500] env[69171]: INFO nova.compute.claims [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.582874] env[69171]: INFO nova.scheduler.client.report [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Deleted allocations for instance ef24399e-4d30-4415-b880-c0b0dccdc461 [ 912.091691] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a705f07a-4225-40b6-99d9-a24b89c19108 tempest-ServerGroupTestJSON-276647374 tempest-ServerGroupTestJSON-276647374-project-member] Lock "ef24399e-4d30-4415-b880-c0b0dccdc461" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 6.470s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.501393] env[69171]: DEBUG nova.scheduler.client.report [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 912.516723] env[69171]: DEBUG nova.scheduler.client.report [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 912.516846] env[69171]: DEBUG nova.compute.provider_tree [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.529531] env[69171]: DEBUG nova.scheduler.client.report [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 912.549211] env[69171]: DEBUG nova.scheduler.client.report [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 912.623524] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9a95eba-a6b2-45df-9f51-fb43c5b0df58 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.631926] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc79b31b-da8f-4577-b70e-6cce5d027615 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.662467] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f5cc50-edf1-49fb-ad61-b061fa9fbb2a {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.670706] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1121caa-4d0e-43c3-b1d5-ea1fb5143ef6 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.684997] env[69171]: DEBUG nova.compute.provider_tree [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 913.207166] env[69171]: ERROR nova.scheduler.client.report [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [req-594cc447-7859-4444-b6e9-413c57d1017b] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-594cc447-7859-4444-b6e9-413c57d1017b"}]} [ 913.207560] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.736s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.208165] env[69171]: ERROR nova.compute.manager [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Failed to build and run instance: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 913.208165] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Traceback (most recent call last): [ 913.208165] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 913.208165] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] yield [ 913.208165] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 913.208165] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] self.set_inventory_for_provider( [ 913.208165] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 913.208165] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-594cc447-7859-4444-b6e9-413c57d1017b"}]} [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] During handling of the above exception, another exception occurred: [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Traceback (most recent call last): [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/compute/manager.py", line 2616, in _build_and_run_instance [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] with self.rt.instance_claim(context, instance, node, allocs, [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 913.208454] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] return f(*args, **kwargs) [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 215, in instance_claim [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] self._update(elevated, cn) [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] self._update_to_placement(context, compute_node, startup) [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] return attempt.get(self._wrap_exception) [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] six.reraise(self.value[0], self.value[1], self.value[2]) [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] raise value [ 913.208673] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] self.reportclient.update_from_provider_tree( [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] with catch_all(pd.uuid): [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] self.gen.throw(typ, value, traceback) [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] raise exception.ResourceProviderSyncFailed() [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 913.208968] env[69171]: ERROR nova.compute.manager [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] [ 913.209215] env[69171]: DEBUG nova.compute.utils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 913.211147] env[69171]: DEBUG nova.compute.manager [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Build of instance 1214f461-59f9-4a21-ae58-ee4803622e8e was re-scheduled: Failed to synchronize the placement service with resource provider information supplied by the compute host. {{(pid=69171) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2481}} [ 913.211549] env[69171]: DEBUG nova.compute.manager [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Unplugging VIFs for instance {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3007}} [ 913.211778] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquiring lock "refresh_cache-1214f461-59f9-4a21-ae58-ee4803622e8e" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.211984] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Acquired lock "refresh_cache-1214f461-59f9-4a21-ae58-ee4803622e8e" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.212238] env[69171]: DEBUG nova.network.neutron [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Building network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2010}} [ 913.579913] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 913.580477] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Cleaning up deleted instances {{(pid=69171) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11229}} [ 913.736876] env[69171]: DEBUG nova.network.neutron [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 913.833514] env[69171]: DEBUG nova.network.neutron [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.084036] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] There are 2 instances to clean {{(pid=69171) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11238}} [ 914.084252] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: fa5997fa-817c-42b6-a7fa-b3b85a109b6d] Instance has had 0 of 5 cleanup attempts {{(pid=69171) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11242}} [ 914.336488] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Releasing lock "refresh_cache-1214f461-59f9-4a21-ae58-ee4803622e8e" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.336730] env[69171]: DEBUG nova.compute.manager [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=69171) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3030}} [ 914.336925] env[69171]: DEBUG nova.compute.manager [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Deallocating network for instance {{(pid=69171) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2290}} [ 914.337112] env[69171]: DEBUG nova.network.neutron [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] deallocate_for_instance() {{(pid=69171) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1803}} [ 914.353538] env[69171]: DEBUG nova.network.neutron [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 914.587753] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: f8b20421-15d3-46f3-a504-172d044ff4d3] Instance has had 0 of 5 cleanup attempts {{(pid=69171) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11242}} [ 914.856737] env[69171]: DEBUG nova.network.neutron [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.092301] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 915.092490] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Cleaning up deleted instances with incomplete migration {{(pid=69171) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11267}} [ 915.359889] env[69171]: INFO nova.compute.manager [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] [instance: 1214f461-59f9-4a21-ae58-ee4803622e8e] Took 1.02 seconds to deallocate network for instance. [ 915.595111] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 916.394909] env[69171]: INFO nova.scheduler.client.report [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Deleted allocations for instance 1214f461-59f9-4a21-ae58-ee4803622e8e [ 916.902647] env[69171]: DEBUG oslo_concurrency.lockutils [None req-568e1cf7-0996-487a-bdd9-7bc3586a7cee tempest-AttachVolumeShelveTestJSON-695487292 tempest-AttachVolumeShelveTestJSON-695487292-project-member] Lock "1214f461-59f9-4a21-ae58-ee4803622e8e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 6.464s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.094067] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.094067] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 918.094067] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Starting heal instance info cache {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9929}} [ 918.615394] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.615520] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquired lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.615614] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Forcefully refreshing network info cache for instance {{(pid=69171) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2004}} [ 919.145658] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Instance cache missing network info. {{(pid=69171) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3323}} [ 919.709052] env[69171]: DEBUG nova.network.neutron [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Updating instance_info_cache with network_info: [] {{(pid=69171) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.212628] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Releasing lock "refresh_cache-8e6a9d73-6705-484c-82e0-70c6a6860211" {{(pid=69171) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.213037] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [instance: 8e6a9d73-6705-484c-82e0-70c6a6860211] Updated the network info_cache for instance {{(pid=69171) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10000}} [ 920.213103] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.213265] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.213415] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.213567] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.213737] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.213884] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.214058] env[69171]: DEBUG nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=69171) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10548}} [ 920.214251] env[69171]: DEBUG oslo_service.periodic_task [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Running periodic task ComputeManager.update_available_resource {{(pid=69171) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.717777] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.718024] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.718198] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.718355] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=69171) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 920.719282] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fc198af-c0da-44a9-b833-21eb9cad2588 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.728656] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2e7db3a-1265-4861-a524-500afa8bb138 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.743130] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638c7d40-4e9d-457a-96e7-c9bc1493a74f {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.749614] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42445679-ae6d-40f9-a1e1-4acf1cb9cd4d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.778047] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180337MB free_disk=0GB free_vcpus=48 pci_devices=None {{(pid=69171) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 920.778208] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.778402] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.805854] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance f5a71cca-1e32-4c3c-9bb8-83e634a864f7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.806122] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 8e6a9d73-6705-484c-82e0-70c6a6860211 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.806157] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 6a29d495-c938-41ad-97b6-1331b6eff589 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.806294] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Instance 4c0f5ce7-ba5b-4000-b8ad-638d0b2fbf03 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=69171) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 921.806464] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 921.806597] env[69171]: DEBUG nova.compute.resource_tracker [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=69171) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 921.824707] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing inventories for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 921.838453] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating ProviderTree inventory for provider 560320dd-0035-4785-a81a-5134ce1f7aaa from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 921.838646] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 1, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 921.849607] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing aggregate associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, aggregates: None {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 921.867178] env[69171]: DEBUG nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Refreshing trait associations for resource provider 560320dd-0035-4785-a81a-5134ce1f7aaa, traits: COMPUTE_NODE,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_NET_ATTACH_INTERFACE {{(pid=69171) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 921.919172] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24d6011-404d-4117-b221-5767b2367876 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.927264] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3baaed-472d-413b-afb3-34f5fdb59f5d {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.958174] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929116a2-1c48-4d06-a66b-b07c61800910 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.965535] env[69171]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd8e710-208d-48f7-a036-787cb85e6af2 {{(pid=69171) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.978444] env[69171]: DEBUG nova.compute.provider_tree [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Updating inventory in ProviderTree for provider 560320dd-0035-4785-a81a-5134ce1f7aaa with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=69171) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 922.498661] env[69171]: ERROR nova.scheduler.client.report [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] [req-6f83cd8d-bfa7-4724-a56c-30cb4f73631a] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 0, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID 560320dd-0035-4785-a81a-5134ce1f7aaa. Got 400: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6f83cd8d-bfa7-4724-a56c-30cb4f73631a"}]} [ 922.498973] env[69171]: DEBUG oslo_concurrency.lockutils [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 1.721s {{(pid=69171) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.499533] env[69171]: ERROR nova.compute.manager [None req-a16a6a55-bb0a-45ff-953f-667aa53cf896 None None] Error updating resources for node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28.: nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 922.499533] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 922.499533] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1402, in catch_all [ 922.499533] env[69171]: ERROR nova.compute.manager yield [ 922.499533] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1498, in update_from_provider_tree [ 922.499533] env[69171]: ERROR nova.compute.manager self.set_inventory_for_provider( [ 922.499533] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1007, in set_inventory_for_provider [ 922.499533] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderUpdateFailed(url=url, error=resp.text) [ 922.499533] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderUpdateFailed: Failed to update resource provider via URL /resource_providers/560320dd-0035-4785-a81a-5134ce1f7aaa/inventories: {"errors": [{"status": 400, "title": "Bad Request", "detail": "The server could not comply with the request since it is either malformed or otherwise incorrect.\n\n JSON does not validate: 0 is less than the minimum of 1 Failed validating 'minimum' in schema['properties']['inventories']['patternProperties']['^[A-Z0-9_]+$']['properties']['max_unit']: {'maximum': 2147483647, 'minimum': 1, 'type': 'integer'} On instance['inventories']['DISK_GB']['max_unit']: 0 ", "code": "placement.undefined_code", "request_id": "req-6f83cd8d-bfa7-4724-a56c-30cb4f73631a"}]} [ 922.499533] env[69171]: ERROR nova.compute.manager [ 922.499533] env[69171]: ERROR nova.compute.manager During handling of the above exception, another exception occurred: [ 922.499533] env[69171]: ERROR nova.compute.manager [ 922.499894] env[69171]: ERROR nova.compute.manager Traceback (most recent call last): [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 10584, in _update_available_resource_for_node [ 922.499894] env[69171]: ERROR nova.compute.manager self.rt.update_available_resource(context, nodename, [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 935, in update_available_resource [ 922.499894] env[69171]: ERROR nova.compute.manager self._update_available_resource(context, resources, startup=startup) [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py", line 412, in inner [ 922.499894] env[69171]: ERROR nova.compute.manager return f(*args, **kwargs) [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1066, in _update_available_resource [ 922.499894] env[69171]: ERROR nova.compute.manager self._update(context, cn, startup=startup) [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1375, in _update [ 922.499894] env[69171]: ERROR nova.compute.manager self._update_to_placement(context, compute_node, startup) [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 56, in wrapped_f [ 922.499894] env[69171]: ERROR nova.compute.manager return Retrying(*dargs, **dkw).call(f, *args, **kw) [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 257, in call [ 922.499894] env[69171]: ERROR nova.compute.manager return attempt.get(self._wrap_exception) [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 301, in get [ 922.499894] env[69171]: ERROR nova.compute.manager six.reraise(self.value[0], self.value[1], self.value[2]) [ 922.499894] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/six.py", line 719, in reraise [ 922.500361] env[69171]: ERROR nova.compute.manager raise value [ 922.500361] env[69171]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/retrying.py", line 251, in call [ 922.500361] env[69171]: ERROR nova.compute.manager attempt = Attempt(fn(*args, **kwargs), attempt_number, False) [ 922.500361] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/resource_tracker.py", line 1360, in _update_to_placement [ 922.500361] env[69171]: ERROR nova.compute.manager self.reportclient.update_from_provider_tree( [ 922.500361] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1497, in update_from_provider_tree [ 922.500361] env[69171]: ERROR nova.compute.manager with catch_all(pd.uuid): [ 922.500361] env[69171]: ERROR nova.compute.manager File "/usr/lib/python3.10/contextlib.py", line 153, in __exit__ [ 922.500361] env[69171]: ERROR nova.compute.manager self.gen.throw(typ, value, traceback) [ 922.500361] env[69171]: ERROR nova.compute.manager File "/opt/stack/nova/nova/scheduler/client/report.py", line 1414, in catch_all [ 922.500361] env[69171]: ERROR nova.compute.manager raise exception.ResourceProviderSyncFailed() [ 922.500361] env[69171]: ERROR nova.compute.manager nova.exception.ResourceProviderSyncFailed: Failed to synchronize the placement service with resource provider information supplied by the compute host. [ 922.500361] env[69171]: ERROR nova.compute.manager